repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/critic.py
applications/ColossalChat/coati/models/critic.py
""" Critic model """ from typing import Optional import torch import torch.nn as nn from coati.models import BaseModel from transformers import PretrainedConfig class Critic(BaseModel): """ Critic model class. Args: pretrained (str): path to pretrained model. config (PretrainedConfig): PretrainedConfig used to initiate the base model. """ def __init__(self, pretrained: str = None, config: Optional[PretrainedConfig] = None, **kwargs) -> None: super().__init__(pretrained=pretrained, config=config, **kwargs) # et last hidden state size with dummy input self.value_head = nn.Linear(self.last_hidden_state_size, 1) def forward(self, input_ids: torch.LongTensor, attention_mask: Optional[torch.Tensor] = None) -> torch.Tensor: outputs = self.model(input_ids, attention_mask=attention_mask) last_hidden_states = outputs["last_hidden_state"] sequence_hidden_states = last_hidden_states[torch.arange(last_hidden_states.size(0)), :].type( self.value_head.weight.dtype ) values = self.value_head(sequence_hidden_states).squeeze(-1) # ensure shape is (B, sequence length) return values def get_input_embeddings(self): return self.model.get_input_embeddings() def get_output_embeddings(self): return self.model.get_output_embeddings()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/utils.py
applications/ColossalChat/coati/models/utils.py
import json import os from typing import Any, Dict, Optional, Union import torch import torch.nn.functional as F def get_model_numel(model: torch.nn.Module) -> int: return sum(p.numel() for p in model.parameters()) def compute_reward( r: Union[torch.Tensor, float], kl_coef: float, log_probs: torch.Tensor, log_probs_base: torch.Tensor, action_mask: Optional[torch.Tensor] = None, reward_eps=5, ) -> torch.Tensor: """ Args: log_probs: [batch_size, response_length] log_probs_base: [batch_size, response_length] action_mask: [batch_size, response_length] r: float Returns: reward: [batch_size, response_length] """ log_ratio = log_probs - log_probs_base # address numerical instability issue kl = -kl_coef * log_ratio * action_mask reward = kl r_clip = torch.clamp(r, -reward_eps, reward_eps) for i in range(action_mask.size(0)): assert action_mask[i].sum() > 0 reward[i, : action_mask[i].sum()] += r_clip[i] reward[i, action_mask[i].sum() :] *= 0 return reward, ((log_ratio * (log_ratio < 10)).exp() - 1 - log_ratio) * action_mask def _log_probs_from_logits(logits: torch.Tensor, labels: torch.Tensor) -> torch.Tensor: """ Compute the log probabilities from logits for the given labels. Args: logits (torch.Tensor): The input logits. labels (torch.Tensor): The target labels. Returns: torch.Tensor: The log probabilities corresponding to the labels. """ log_probs = F.log_softmax(logits, dim=-1) per_label_logps = log_probs.gather(dim=-1, index=labels.unsqueeze(-1)) return per_label_logps.squeeze(-1) def calc_action_log_probs(logits: torch.Tensor, sequences: torch.LongTensor, num_actions: int) -> torch.Tensor: """Calculate action log probs. Args: output (torch.Tensor): Output tensor of Actor.forward.logits. sequences (torch.LongTensor): Input sequences. num_actions (int): Number of actions. Returns: torch.Tensor: Action log probs. """ log_probs = _log_probs_from_logits(logits[:, :-1, :], sequences[:, 1:]) return log_probs[:, -num_actions:] def masked_mean(tensor: torch.Tensor, mask: torch.Tensor, dim: int = 1) -> torch.Tensor: """ Compute the masked mean of a tensor along a specified dimension. Args: tensor (torch.Tensor): The input tensor. mask (torch.Tensor): The mask tensor with the same shape as the input tensor. dim (int, optional): The dimension along which to compute the mean. Default is 1. Returns: torch.Tensor: The masked mean tensor. """ tensor = tensor * mask tensor = tensor.sum(dim=dim) mask_sum = mask.sum(dim=dim) mean = tensor / (mask_sum + 1e-8) return mean def calc_masked_log_probs( logits: torch.Tensor, sequences: torch.LongTensor, mask: torch.Tensor, length_normalization: bool = False ) -> torch.Tensor: """ Calculate the masked log probabilities for a given sequence of logits. Args: logits (torch.Tensor): The input logits tensor of shape (batch_size, sequence_length, vocab_size). sequences (torch.LongTensor): The input sequence tensor of shape (batch_size, sequence_length). mask (torch.Tensor): The mask tensor of shape (batch_size, sequence_length). Returns: torch.Tensor: The masked log probabilities tensor of shape (batch_size, sequence_length - 1). """ # logits are probabilities of the next token, so we shift them to the left by one log_probs = _log_probs_from_logits(logits[:, :-1, :], sequences[:, 1:]) if not length_normalization: return log_probs * mask else: return log_probs * mask / (mask.sum(dim=-1, keepdim=True) + 0.01) def load_json(file_path: Union[str, os.PathLike]) -> Dict[str, Any]: """ Load file in JSON format """ with open(file=file_path, mode="r", encoding="utf-8") as fp: return json.load(fp) def save_json(data: Dict[str, Any], file_path: Union[str, os.PathLike]) -> None: """ Save as JSON format """ with open(file=file_path, mode="w", encoding="utf-8") as fp: json.dump(data, fp=fp, ensure_ascii=False, indent=4) def disable_dropout(model: torch.nn.Module): """ Disables dropout in a PyTorch model. This is used in PPO Training Args: model (torch.nn.Module): The PyTorch model. Returns: None """ if model is not None: for module in model.modules(): if isinstance(module, torch.nn.Dropout): module.p = 0.0 def repad_to_left(tensor, tokenizer): repadded_input_ids = [] max_non_padded_seq_len = 0 for i in range(tensor.size(0)): non_pad_indices = (tensor[i] != tokenizer.pad_token_id).nonzero(as_tuple=True)[0] start, end = non_pad_indices.min(), non_pad_indices.max() repadded_input_ids.append(tensor[i][start : end + 1]) max_non_padded_seq_len = max(max_non_padded_seq_len, repadded_input_ids[-1].size(0)) repadded_input_ids = [ F.pad(t, (max_non_padded_seq_len - t.size(0), 0), value=tokenizer.pad_token_id) for t in repadded_input_ids ] return torch.stack(repadded_input_ids)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/loss.py
applications/ColossalChat/coati/models/loss.py
""" loss functions """ from typing import Optional, Tuple import torch import torch.distributed as dist import torch.nn as nn from .utils import masked_mean class GPTLMLoss(nn.Module): """ GPT Language Model Loss """ def __init__(self): super().__init__() # NOTE: default ignore_index is -100, which is equal to IGNORE_INDEX in sft_dataset.py self.loss = nn.CrossEntropyLoss() def forward(self, logits: torch.Tensor, labels: torch.Tensor) -> torch.Tensor: shift_logits = logits[..., :-1, :].contiguous() shift_labels = labels[..., 1:].contiguous() # Flatten the tokens return self.loss(shift_logits.view(-1, shift_logits.size(-1)), shift_labels.view(-1)) class PolicyLoss(nn.Module): """ Policy Loss for PPO """ def __init__(self, clip_eps: float = 0.2, skip_threshold: float = 20.0) -> None: super().__init__() self.clip_eps = clip_eps self.skip_threshold = skip_threshold def forward( self, log_probs: torch.Tensor, old_log_probs: torch.Tensor, advantages: torch.Tensor, action_mask: Optional[torch.Tensor] = None, ) -> torch.Tensor: skip = False if action_mask is None: ratio_ = (log_probs - old_log_probs).exp() else: ratio_ = ((log_probs - old_log_probs) * action_mask).exp() # note that if dropout is disabled (recommanded), ratio will always be 1. if ratio_.mean() > self.skip_threshold: skip = True ratio = ratio_.clamp(0.0, 10.0) surr1 = ratio * advantages surr2 = ratio.clamp(1 - self.clip_eps, 1 + self.clip_eps) * advantages loss = -torch.min(surr1, surr2) if action_mask is not None: loss = masked_mean(loss, action_mask) else: loss = loss.mean(dim=1) loss = loss.mean() return loss, skip, ratio_.max() class ValueLoss(nn.Module): """ Value Loss for PPO """ def __init__(self, clip_eps: float = 0.2) -> None: super().__init__() self.clip_eps = clip_eps def forward( self, values: torch.Tensor, old_values: torch.Tensor, advantage: torch.Tensor, action_mask: Optional[torch.Tensor] = None, ) -> torch.Tensor: returns = advantage + old_values values_clipped = old_values + (values - old_values).clamp(-self.clip_eps, self.clip_eps) surr1 = (values_clipped - returns) ** 2 surr2 = (values - returns) ** 2 if action_mask is not None: loss = torch.sum(torch.max(surr1, surr2) / torch.sum(action_mask) * action_mask) else: loss = torch.mean(torch.max(surr1, surr2)) return 0.5 * loss class DpoLoss(nn.Module): """ Dpo loss Details: https://arxiv.org/pdf/2305.18290.pdf SimPO loss: Details: https://arxiv.org/pdf/2405.14734.pdf """ def __init__(self, beta: float = 0.1, gamma: float = 0.0): """ Args: beta: The temperature parameter in the DPO paper. gamma: The margin parameter in the SimPO paper. length_normalization: Whether to normalize the loss by the length of chosen and rejected responses. Refer to the length normalization in the SimPO paper """ super().__init__() self.beta = beta self.gamma = gamma def forward( self, logprob_actor_chosen: torch.Tensor, logprob_actor_reject: torch.Tensor, logprob_ref_chosen: torch.Tensor, logprob_ref_reject: torch.Tensor, chosen_mask: torch.Tensor, reject_mask: torch.Tensor, ) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]: """Compute the DPO/SimPO loss for a batch of policy and reference model log probabilities. # adapted from https://github.com/huggingface/trl/blob/main/trl/trainer/dpo_trainer.py#L328 Args: logprob_actor_chosen: Log probabilities of the policy model for the chosen responses. Shape: (batch_size,) logprob_actor_reject: Log probabilities of the policy model for the rejected responses. Shape: (batch_size,) logprob_ref_chosen: Log probabilities of the reference model for the chosen responses. Shape: (batch_size,) logprob_ref_reject: Log probabilities of the reference model for the rejected responses. Shape: (batch_size,) chosen_mask: Mask tensor indicating which responses were chosen. Shape: (batch_size,) reject_mask: Mask tensor indicating which responses were rejected. Shape: (batch_size,) Returns: A tuple of three tensors: (losses, chosen_rewards, rejected_rewards). The losses tensor contains the DPO loss for each example in the batch. The chosen_rewards and rejected_rewards tensors contain the rewards for the chosen and rejected responses, respectively. """ logprob_actor_chosen = logprob_actor_chosen * chosen_mask logprob_actor_reject = logprob_actor_reject * reject_mask if logprob_ref_chosen is not None and logprob_ref_reject is not None: logprob_ref_chosen = logprob_ref_chosen * chosen_mask logprob_ref_reject = logprob_ref_reject * reject_mask if len(logprob_ref_chosen.shape) == 2: ref_logratios = logprob_ref_chosen.sum(-1) - logprob_ref_reject.sum(-1) else: ref_logratios = logprob_ref_chosen - logprob_ref_reject else: # If no reference model is provided ref_logratios = 0.0 pi_logratios = logprob_actor_chosen.sum(-1) - logprob_actor_reject.sum(-1) logits = pi_logratios - ref_logratios - self.gamma / self.beta losses = -torch.nn.functional.logsigmoid(self.beta * logits) loss = losses.mean() # Calculate rewards for logging if logprob_ref_chosen is not None: chosen_rewards = self.beta * (logprob_actor_chosen.sum(-1) - logprob_ref_chosen.sum(-1)).detach() else: chosen_rewards = self.beta * logprob_actor_chosen.sum(-1).detach() if logprob_ref_reject is not None: rejected_rewards = self.beta * (logprob_actor_reject.sum(-1) - logprob_ref_reject.sum(-1)).detach() else: rejected_rewards = self.beta * logprob_actor_reject.sum(-1).detach() return loss, chosen_rewards, rejected_rewards class LogSigLoss(nn.Module): """ Pairwise Loss for Reward Model Details: https://arxiv.org/abs/2203.02155 """ def forward(self, chosen_reward: torch.Tensor, reject_reward: torch.Tensor) -> torch.Tensor: return -torch.nn.functional.logsigmoid(chosen_reward - reject_reward).mean() class LogExpLoss(nn.Module): """ Pairwise Loss for Reward Model Details: https://arxiv.org/abs/2204.05862 """ def forward(self, chosen_reward: torch.Tensor, reject_reward: torch.Tensor) -> torch.Tensor: loss = torch.log(1 + torch.exp(reject_reward - chosen_reward)).mean() return loss class OddsRatioLoss(nn.Module): """ Odds Ratio Loss in ORPO Details: https://arxiv.org/pdf/2403.07691 """ def forward( self, chosen_logp: torch.Tensor, reject_logp: torch.Tensor, chosen_loss_mask: torch.Tensor, reject_loss_mask: torch.Tensor, ) -> torch.Tensor: chosen_logp = chosen_logp.to(dtype=torch.float32) reject_logp = reject_logp.to(dtype=torch.float32) chosen_odds = chosen_logp - torch.log(-torch.exp(chosen_logp) + 1.0001) chosen_odds_masked = torch.sum(chosen_odds * chosen_loss_mask.float()) / torch.sum(chosen_loss_mask) reject_odds = reject_logp - torch.log(-torch.exp(reject_logp) + 1.0001) reject_odds_masked = torch.sum(reject_odds * reject_loss_mask.float()) / torch.sum(reject_loss_mask) log_odds_ratio = chosen_odds_masked - reject_odds_masked ratio = torch.log(torch.nn.functional.sigmoid(log_odds_ratio)) return ratio.to(dtype=torch.bfloat16), log_odds_ratio class KTOLoss(nn.Module): def __init__(self, beta: float = 0.1, desirable_weight: float = 1.0, undesirable_weight: float = 1.0): """ Args: beta: The temperature parameter in the KTO paper. desirable_weight: The weight for the desirable responses. undesirable_weight: The weight for the undesirable """ super().__init__() self.beta = beta self.desirable_weight = desirable_weight self.undesirable_weight = undesirable_weight def forward( self, chosen_logps: torch.Tensor, rejected_logps: torch.Tensor, kl_logps: torch.Tensor, ref_chosen_logps: torch.Tensor, ref_rejected_logps: torch.Tensor, ref_kl_logps: torch.Tensor, ): """ Reference: https://github.com/huggingface/trl/blob/a2adfb836a90d1e37b1253ab43dace05f1241e04/trl/trainer/kto_trainer.py#L585 Compute the KTO loss for a batch of policy and reference model log probabilities. Args: chosen_logps: Log probabilities of the policy model for the chosen responses. Shape: (batch_size,) rejected_logps: Log probabilities of the policy model for the rejected responses. Shape: (batch_size,) kl_logps: KL divergence of the policy model. Shape: (batch_size,) ref_chosen_logps: Log probabilities of the reference model for the chosen responses. Shape: (batch_size,) ref_rejected_logps: Log probabilities of the reference model for the rejected responses. Shape: (batch_size,) ref_kl_logps: KL divergence of the reference model. Shape: (batch_size,) beta: The temperature parameter in the DPO paper. desirable_weight: The weight for the desirable responses. undesirable_weight: The weight for the undesirable responses. Refer to the KTO paper for details about hyperparameters https://arxiv.org/pdf/2402.01306 """ kl = (kl_logps - ref_kl_logps).mean().detach() # all gather dist.all_reduce(kl, op=dist.ReduceOp.SUM) kl = (kl / dist.get_world_size()).clamp(min=0) if chosen_logps.shape[0] != 0 and ref_chosen_logps.shape[0] != 0: chosen_logratios = chosen_logps - ref_chosen_logps chosen_losses = 1 - nn.functional.sigmoid(self.beta * (chosen_logratios - kl)) chosen_rewards = self.beta * chosen_logratios.detach() else: chosen_losses = torch.Tensor([]).to(kl_logps.device) chosen_rewards = torch.Tensor([]).to(kl_logps.device) if rejected_logps.shape[0] != 0 and ref_rejected_logps.shape[0] != 0: rejected_logratios = rejected_logps - ref_rejected_logps rejected_losses = 1 - nn.functional.sigmoid(self.beta * (kl - rejected_logratios)) rejected_rewards = self.beta * rejected_logratios.detach() else: rejected_losses = torch.Tensor([]).to(kl_logps.device) rejected_rewards = torch.Tensor([]).to(kl_logps.device) losses = torch.cat((self.desirable_weight * chosen_losses, self.undesirable_weight * rejected_losses), 0).mean() return losses, chosen_rewards, rejected_rewards, kl
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/__init__.py
applications/ColossalChat/coati/models/__init__.py
from .base import BaseModel from .critic import Critic from .generation import generate, generate_streaming, prepare_inputs_fn, update_model_kwargs_fn from .lora import LoraConfig, convert_to_lora_module, lora_manager from .loss import DpoLoss, KTOLoss, LogExpLoss, LogSigLoss, PolicyLoss, ValueLoss from .reward_model import RewardModel from .rlvr_reward_model import RLVRRewardModel from .utils import disable_dropout __all__ = [ "BaseModel", "Critic", "RewardModel", "RLVRRewardModel", "PolicyLoss", "ValueLoss", "LogSigLoss", "LogExpLoss", "LoraConfig", "lora_manager", "convert_to_lora_module", "DpoLoss", "KTOLoss" "generate", "generate_streaming", "disable_dropout", "update_model_kwargs_fn", "prepare_inputs_fn", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/base.py
applications/ColossalChat/coati/models/base.py
""" Base class for critic and reward model """ from typing import Optional import torch import torch.nn as nn from transformers import AutoModel, PretrainedConfig class BaseModel(nn.Module): """ Actor model base class. Args: pretrained (str): path to pretrained model. config (PretrainedConfig): PretrainedConfig used to initiate the base model. **kwargs: all other kwargs as in AutoModel.from_pretrained """ def __init__(self, pretrained: str = None, config: Optional[PretrainedConfig] = None, **kwargs) -> None: super().__init__() if pretrained is not None: if config is not None: # initialize with config and load weights from pretrained self.model = AutoModel.from_pretrained(pretrained, config=config, **kwargs) else: # initialize with pretrained self.model = AutoModel.from_pretrained(pretrained, **kwargs) elif config is not None: # initialize with config self.model = AutoModel.from_config(config, **kwargs) else: raise ValueError("Either pretrained or config must be provided.") self.config = self.model.config # create dummy input to get the size of the last hidden state if "use_flash_attention_2" in kwargs: self.model = self.model.cuda() dummy_input = torch.zeros((1, 1), dtype=torch.long).to(self.model.device) out = self.model(dummy_input) self.last_hidden_state_size = out.last_hidden_state.shape[-1] self.model = self.model.cpu() def resize_token_embeddings(self, *args, **kwargs): """ Resize the token embeddings of the model. Args: *args: Variable length argument list. **kwargs: Arbitrary keyword arguments. Returns: The resized token embeddings. """ return self.model.resize_token_embeddings(*args, **kwargs)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/generation.py
applications/ColossalChat/coati/models/generation.py
import copy from typing import Any, Callable, List, Optional import torch import torch.distributed as dist from transformers import PreTrainedTokenizer try: from transformers.generation_logits_process import ( LogitsProcessorList, TemperatureLogitsWarper, TopKLogitsWarper, TopPLogitsWarper, ) except ImportError: from transformers.generation import LogitsProcessorList, TemperatureLogitsWarper, TopKLogitsWarper, TopPLogitsWarper def _prepare_logits_processor( top_k: Optional[int] = None, top_p: Optional[float] = None, temperature: Optional[float] = None ) -> LogitsProcessorList: """ Prepare the logits processor list based on the given parameters. Args: top_k (Optional[int]): The number of highest probability logits to keep for each token. top_p (Optional[float]): The cumulative probability threshold for selecting tokens. temperature (Optional[float]): The temperature value to apply to the logits. Returns: LogitsProcessorList: The list of logits processors. """ processor_list = LogitsProcessorList() if temperature is not None and temperature != 1.0: processor_list.append(TemperatureLogitsWarper(temperature)) if top_k is not None and top_k != 0: processor_list.append(TopKLogitsWarper(top_k)) if top_p is not None and top_p < 1.0: processor_list.append(TopPLogitsWarper(top_p)) return processor_list def _is_sequence_finished(unfinished_sequences: torch.Tensor) -> bool: """ Check if the sequence generation is finished. Args: unfinished_sequences (torch.Tensor): Tensor indicating the unfinished sequences. Returns: bool: True if all sequences are finished, False otherwise. """ if dist.is_initialized() and dist.get_world_size() > 1: # consider DP unfinished_sequences = unfinished_sequences.clone() dist.all_reduce(unfinished_sequences) return unfinished_sequences.max() == 0 def update_model_kwargs_fn(outputs: dict, new_mask, **model_kwargs) -> dict: """ Update the model keyword arguments based on the outputs and new mask. Args: outputs (dict): The outputs from the model. new_mask: The new attention mask. **model_kwargs: Additional model keyword arguments. Returns: dict: The updated model keyword arguments. """ if "past_key_values" in outputs: model_kwargs["past_key_values"] = outputs["past_key_values"] else: model_kwargs["past_key_values"] = None # update token_type_ids with last value if "token_type_ids" in model_kwargs: token_type_ids = model_kwargs["token_type_ids"] model_kwargs["token_type_ids"] = torch.cat([token_type_ids, token_type_ids[:, -1].unsqueeze(-1)], dim=-1) # update attention mask if "attention_mask" in model_kwargs: attention_mask = model_kwargs["attention_mask"] model_kwargs["attention_mask"] = torch.cat([attention_mask, new_mask], dim=-1) return model_kwargs def prepare_inputs_fn(input_ids: torch.Tensor, **model_kwargs) -> dict: model_kwargs["input_ids"] = input_ids return model_kwargs def _sample( model: Any, tokenizer: Any, input_ids: torch.Tensor, max_length: int, early_stopping: bool = True, eos_token_id: Optional[int] = None, pad_token_id: Optional[int] = None, stop_token_ids: Optional[List[int]] = None, top_k: Optional[int] = None, top_p: Optional[float] = None, temperature: Optional[float] = None, max_new_tokens: int = None, prepare_inputs_fn: Optional[Callable[[torch.Tensor, Any], dict]] = None, update_model_kwargs_fn: Optional[Callable[[dict, Any], dict]] = None, stream_interval: int = 2, **model_kwargs, ) -> torch.Tensor: """ Generates new tokens using the given model and input_ids. Args: model (Any): The model used for token generation. input_ids (torch.Tensor): The input tensor containing the initial tokens. max_length (int): The maximum length of the generated tokens. early_stopping (bool, optional): Whether to stop generating tokens early if all sequences are finished. Defaults to True. eos_token_id (int, optional): The ID of the end-of-sequence token. Defaults to None. pad_token_id (int, optional): The ID of the padding token. Defaults to None. stop_token_ids (List[int], optional): A list of token IDs that, if encountered, will stop the generation process. Defaults to None. top_k (int, optional): The number of top-k tokens to consider during sampling. Defaults to None. top_p (float, optional): The cumulative probability threshold for top-p sampling. Defaults to None. temperature (float, optional): The temperature value for token sampling. Defaults to None. max_new_tokens (int, optional): The maximum number of new tokens to generate. Defaults to None. prepare_inputs_fn (Callable[[torch.Tensor, Any], dict], optional): A function to prepare the model inputs. Defaults to None. update_model_kwargs_fn (Callable[[dict, Any], dict], optional): A function to update the model kwargs. Defaults to None. stream_interval (int, optional): The interval for streaming generation. Defaults to 2. **model_kwargs: Additional keyword arguments for the model. Returns: torch.Tensor: The tensor containing the generated tokens. """ context_length = input_ids.size(1) if max_new_tokens is None: max_new_tokens = max_length - context_length if context_length + max_new_tokens > max_length or max_new_tokens == 0: print("Exeeded length limitation") return input_ids logits_processor = _prepare_logits_processor(top_k, top_p, temperature) unfinished_sequences = input_ids.new(input_ids.shape[0]).fill_(1) past = None for i in range(context_length, context_length + max_new_tokens): # Calculate attention mask if "attention_mask" not in model_kwargs: model_kwargs["attention_mask"] = input_ids.ne(pad_token_id) model_inputs = ( prepare_inputs_fn(input_ids, past=past, **model_kwargs) if prepare_inputs_fn is not None else {"input_ids": input_ids, "attention_mask": input_ids.ne(pad_token_id)} ) outputs = model(**model_inputs) if "past_key_values" in outputs: past = outputs.past_key_values elif "mems" in outputs: past = outputs.mems # NOTE: this is correct only in left padding mode next_token_logits = outputs["logits"][:, -1, :] next_token_logits = logits_processor(input_ids, next_token_logits) # Sample probs = torch.softmax(next_token_logits, dim=-1, dtype=torch.float) next_tokens = torch.multinomial(probs, num_samples=1).squeeze(1) # Finished sentences should have their next token be a padding token if eos_token_id is not None: assert pad_token_id is not None, "If `eos_token_id` is defined, make sure that `pad_token_id` is defined." next_tokens = next_tokens * unfinished_sequences + pad_token_id * (1 - unfinished_sequences) # Update generated ids, model inputs for next step input_ids = torch.cat([input_ids, next_tokens[:, None]], dim=-1) if update_model_kwargs_fn is not None: model_kwargs = update_model_kwargs_fn(outputs, model_kwargs) # If eos_token was found in one sentence, set sentence to finished if eos_token_id is not None: unfinished_sequences = unfinished_sequences.mul((next_tokens != eos_token_id).long()) if stop_token_ids is not None: # If the last len(stop_token_ids) tokens of input_ids are equal to stop_token_ids, set sentence to finished. for stop_token_id in stop_token_ids: tokens_to_check = input_ids[:, -len(stop_token_id) :] unfinished_sequences = unfinished_sequences.mul( torch.any(tokens_to_check != torch.LongTensor(stop_token_id).to(input_ids.device), dim=1).long() ) # Stop when each sentence is finished if early_stopping=True if (early_stopping and _is_sequence_finished(unfinished_sequences)) or i == context_length + max_new_tokens - 1: return input_ids @torch.inference_mode() def generate( model: Any, input_ids: torch.Tensor, tokenizer: PreTrainedTokenizer, max_length: int, num_beams: int = 1, do_sample: bool = True, early_stopping: bool = True, top_k: Optional[int] = None, top_p: Optional[float] = None, temperature: Optional[float] = None, prepare_inputs_fn: Optional[Callable[[torch.Tensor, Any], dict]] = None, update_model_kwargs_fn: Optional[Callable[[dict, Any], dict]] = None, **model_kwargs, ) -> torch.Tensor: """Generate token sequence. The returned sequence is input_ids + generated_tokens. Args: model (nn.Module): model input_ids (torch.Tensor): input sequence max_length (int): max length of the returned sequence num_beams (int, optional): number of beams. Defaults to 1. do_sample (bool, optional): whether to do sample. Defaults to True. early_stopping (bool, optional): if True, the sequence length may be smaller than max_length due to finding eos. Defaults to False. top_k (Optional[int], optional): the number of highest probability vocabulary tokens to keep for top-k-filtering. Defaults to None. top_p (Optional[float], optional): If set to float < 1, only the smallest set of most probable tokens with probabilities that add up to top_p or higher are kept for generation. Defaults to None. temperature (Optional[float], optional): The value used to module the next token probabilities. Defaults to None. prepare_inputs_fn (Optional[Callable[[torch.Tensor, Any], dict]], optional): Function to preprocess model inputs. Arguments of this function should be input_ids and model_kwargs. Defaults to None. update_model_kwargs_fn (Optional[Callable[[dict, Any], dict]], optional): Function to update model_kwargs based on outputs. Arguments of this function should be outputs and model_kwargs. Defaults to None. """ assert tokenizer.padding_side == "left", "Current generation only supports left padding." is_greedy_gen_mode = (num_beams == 1) and do_sample is False is_sample_gen_mode = (num_beams == 1) and do_sample is True is_beam_gen_mode = (num_beams > 1) and do_sample is False if is_greedy_gen_mode: raise NotImplementedError elif is_sample_gen_mode: # Run sample generation_kwargs = copy.deepcopy(model_kwargs) res = _sample( model, tokenizer, input_ids, max_length, early_stopping=early_stopping, eos_token_id=tokenizer.eos_token_id, pad_token_id=tokenizer.pad_token_id, top_k=top_k, top_p=top_p, temperature=temperature, prepare_inputs_fn=prepare_inputs_fn, update_model_kwargs_fn=update_model_kwargs_fn, **generation_kwargs, ) del generation_kwargs return res elif is_beam_gen_mode: raise NotImplementedError else: raise ValueError("Unsupported generation mode") def _sample_streaming( model: Any, input_ids: torch.Tensor, max_length: int, early_stopping: bool = False, eos_token_id: Optional[int] = None, pad_token_id: Optional[int] = None, stop_token_ids: Optional[List[int]] = None, top_k: Optional[int] = None, top_p: Optional[float] = None, temperature: Optional[float] = None, max_new_tokens: int = None, prepare_inputs_fn: Optional[Callable[[torch.Tensor, Any], dict]] = None, update_model_kwargs_fn: Optional[Callable[[dict, Any], dict]] = None, stream_interval: int = 2, **model_kwargs, ) -> torch.Tensor: """ Generates new tokens using a streaming approach. Args: model (Any): The model used for token generation. input_ids (torch.Tensor): The input tensor containing the initial tokens. max_length (int): The maximum length of the generated sequence. early_stopping (bool, optional): Whether to stop generating tokens for a sequence if it is finished. Defaults to False. eos_token_id (int, optional): The ID of the end-of-sequence token. Defaults to None. pad_token_id (int, optional): The ID of the padding token. Defaults to None. stop_token_ids (List[int], optional): A list of token IDs that, if encountered, will mark the sequence as finished. Defaults to None. top_k (int, optional): The number of top-k tokens to consider during sampling. Defaults to None. top_p (float, optional): The cumulative probability threshold for top-p sampling. Defaults to None. temperature (float, optional): The temperature value for sampling. Defaults to None. max_new_tokens (int, optional): The maximum number of new tokens to generate. Defaults to None. prepare_inputs_fn (Callable[[torch.Tensor, Any], dict], optional): A function to prepare the model inputs. Defaults to None. update_model_kwargs_fn (Callable[[dict, Any], dict], optional): A function to update the model keyword arguments. Defaults to None. stream_interval (int, optional): The interval at which to yield the generated tokens. Defaults to 2. **model_kwargs: Additional keyword arguments to be passed to the model. Yields: torch.Tensor: The generated tokens at each step. Returns: torch.Tensor: The final generated tokens. """ context_length = input_ids.size(1) if max_new_tokens is None: max_new_tokens = max_length - context_length if context_length + max_new_tokens > max_length or max_new_tokens == 0: return input_ids logits_processor = _prepare_logits_processor(top_k, top_p, temperature) unfinished_sequences = input_ids.new(input_ids.shape[0]).fill_(1) past = None for i in range(context_length, context_length + max_new_tokens): # calculate attention mask if "attention_mask" not in model_kwargs: model_kwargs["attention_mask"] = input_ids.ne(pad_token_id) model_inputs = ( prepare_inputs_fn(input_ids, past=past, **model_kwargs) if prepare_inputs_fn is not None else {"input_ids": input_ids, "attention_mask": input_ids.ne(pad_token_id)} ) outputs = model(**model_inputs) if "past_key_values" in outputs: past = outputs.past_key_values elif "mems" in outputs: past = outputs.mems # NOTE: this is correct only in left padding mode next_token_logits = outputs["logits"][:, -1, :] next_token_logits = logits_processor(input_ids, next_token_logits) # sample probs = torch.softmax(next_token_logits, dim=-1, dtype=torch.float) next_tokens = torch.multinomial(probs, num_samples=1).squeeze(1) # finished sentences should have their next token be a padding token if eos_token_id is not None: assert pad_token_id is not None, "If `eos_token_id` is defined, make sure that `pad_token_id` is defined." next_tokens = next_tokens * unfinished_sequences + pad_token_id * (1 - unfinished_sequences) # update generated ids, model inputs for next step input_ids = torch.cat([input_ids, next_tokens[:, None]], dim=-1) if update_model_kwargs_fn is not None: model_kwargs = update_model_kwargs_fn(outputs, model_kwargs) # if eos_token was found in one sentence, set sentence to finished if eos_token_id is not None: unfinished_sequences = unfinished_sequences.mul((next_tokens != eos_token_id).long()) if stop_token_ids is not None: tokens_to_check = input_ids[:, -len(stop_token_ids) :] if isinstance(stop_token_ids[0], int): # If the last len(stop_token_ids) tokens of input_ids are equal to stop_token_ids, set sentence to finished. unfinished_sequences = unfinished_sequences.mul( torch.any(tokens_to_check != torch.LongTensor(stop_token_ids).to(input_ids.device), dim=1).long() ) else: for stop_token_id in stop_token_ids: unfinished_sequences = unfinished_sequences.mul( torch.any(tokens_to_check != torch.LongTensor(stop_token_id).to(input_ids.device), dim=1).long() ) # Stop when each sentence is finished if early_stopping=True if ( (early_stopping and _is_sequence_finished(unfinished_sequences)) or (i - context_length) % stream_interval == 0 or i == context_length + max_new_tokens - 1 ): yield input_ids if early_stopping and _is_sequence_finished(unfinished_sequences): break @torch.inference_mode() def generate_streaming( model: Any, input_ids: torch.Tensor, tokenizer: PreTrainedTokenizer, max_length: int, num_beams: int = 1, do_sample: bool = True, early_stopping: bool = False, top_k: Optional[int] = None, top_p: Optional[float] = None, temperature: Optional[float] = None, prepare_inputs_fn: Optional[Callable[[torch.Tensor, Any], dict]] = None, update_model_kwargs_fn: Optional[Callable[[dict, Any], dict]] = None, **model_kwargs, ): """Generate token sequence. The returned sequence is input_ids + generated_tokens. Args: model (nn.Module): model input_ids (torch.Tensor): input sequence max_length (int): max length of the returned sequence num_beams (int, optional): number of beams. Defaults to 1. do_sample (bool, optional): whether to do sample. Defaults to True. early_stopping (bool, optional): if True, the sequence length may be smaller than max_length due to finding eos. Defaults to False. top_k (Optional[int], optional): the number of highest probability vocabulary tokens to keep for top-k-filtering. Defaults to None. top_p (Optional[float], optional): If set to float < 1, only the smallest set of most probable tokens with probabilities that add up to top_p or higher are kept for generation. Defaults to None. temperature (Optional[float], optional): The value used to module the next token probabilities. Defaults to None. prepare_inputs_fn (Optional[Callable[[torch.Tensor, Any], dict]], optional): Function to preprocess model inputs. Arguments of this function should be input_ids and model_kwargs. Defaults to None. update_model_kwargs_fn (Optional[Callable[[dict, Any], dict]], optional): Function to update model_kwargs based on outputs. Arguments of this function should be outputs and model_kwargs. Defaults to None. """ assert tokenizer.padding_side == "left", "Current generation only supports left padding." is_greedy_gen_mode = (num_beams == 1) and do_sample is False is_sample_gen_mode = (num_beams == 1) and do_sample is True is_beam_gen_mode = (num_beams > 1) and do_sample is False if is_greedy_gen_mode: # run greedy search raise NotImplementedError elif is_sample_gen_mode: # run sample for res in _sample_streaming( model, input_ids, max_length, early_stopping=early_stopping, eos_token_id=tokenizer.eos_token_id, pad_token_id=tokenizer.pad_token_id, top_k=top_k, top_p=top_p, temperature=temperature, prepare_inputs_fn=prepare_inputs_fn, update_model_kwargs_fn=update_model_kwargs_fn, **model_kwargs, ): yield res elif is_beam_gen_mode: raise NotImplementedError else: raise ValueError("Unsupported generation mode")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/models/rlvr_reward_model.py
applications/ColossalChat/coati/models/rlvr_reward_model.py
""" reward model """ from typing import Callable, List, Optional import torch class RLVRRewardModel: """ RLVRReward model class. Support varifiable reward. Args: reward_fn_list List: list of reward functions **kwargs: all other kwargs as in reward functions """ def __init__(self, reward_fn_list: List[Callable], **kwargs) -> None: self.reward_fn_list = reward_fn_list self.kwargs = kwargs def __call__( self, input_ids: torch.LongTensor, attention_mask: Optional[torch.Tensor] = None, response_start: List = None, response_end: List = None, gt_answer: List = None, ) -> torch.Tensor: # apply varifiable reward bs = input_ids.size(0) rewards = torch.zeros(bs, device=input_ids.device) for i in range(bs): for reward_fn in self.reward_fn_list: rewards[i] += reward_fn( input_ids[i], attention_mask[i], response_start=response_start[i], response_end=response_end[i], gt_answer=gt_answer[i], **self.kwargs, ) return rewards def to(self, device): return self def eval(self): return self
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/consumer.py
applications/ColossalChat/coati/distributed/consumer.py
from contextlib import nullcontext from typing import Any, Dict, Optional import ray import ray.util.collective as cc import torch import torch.distributed as dist from coati.distributed.profiling_utils import CustomProfiler from coati.utils import save_checkpoint from tqdm import tqdm from transformers import AutoModelForCausalLM from colossalai.booster import Booster from colossalai.booster.plugin import HybridParallelPlugin from colossalai.cluster import DistCoordinator from colossalai.initialize import launch from colossalai.nn.optimizer import HybridAdam from colossalai.utils import get_current_device from .comm import ray_broadcast_tensor_dict from .utils import bind_batch, post_recv, unbind_batch class BaseConsumer: def __init__( self, num_producers: int, num_episodes: int, rank: int, world_size: int, master_addr: str, master_port: int, num_update_per_episode: int, num_recv_per_update: int, batch_size: int, model_config: Dict[str, Any], plugin_config: Dict[str, Any], minibatch_size: int = 1, save_interval: int = 100, save_dir: str = "./model", enable_profiling: bool = False, n_behind: int = 0, ): self.num_producers = num_producers self.num_episodes = num_episodes self.rank = rank self.world_size = world_size self.master_addr = master_addr self.master_port = master_port self.num_update_per_episode = num_update_per_episode self.num_recv_per_update = num_recv_per_update self.batch_size = batch_size self.minibatch_size = minibatch_size self.save_interval = save_interval self.save_dir = save_dir self.enable_profiling = enable_profiling assert batch_size % minibatch_size == 0, "batch_size should be divisible by microbatch_size" self.num_microbatches = batch_size // minibatch_size self.checkpoint_path = model_config.pop("checkpoint_path", None) self.model_config = model_config self.plugin_config = plugin_config self.device = get_current_device() self.lr_scheduler = None self.n_behind = n_behind self.total_prompt_trained = 0 # for setting start index when resume training def setup(self) -> None: launch(self.rank, self.world_size, self.master_addr, self.master_port, local_rank=0) self.coordinator = DistCoordinator() plugin_config = dict(tp_size=1, pp_size=1, precision="bf16", zero_stage=2) if ( self.plugin_config.get("pp_size", 1) > 1 and "num_microbatches" not in self.plugin_config and "microbatch_size" not in self.plugin_config ): plugin_config["microbatch_size"] = max(1, self.minibatch_size // plugin_config.get("pp_size", 1)) plugin_config.update(self.plugin_config) self.plugin = HybridParallelPlugin(**plugin_config) self.booster = Booster(plugin=self.plugin) self.dp_rank = dist.get_rank(self.plugin.dp_group) self.tp_rank = dist.get_rank(self.plugin.tp_group) self.pp_rank = dist.get_rank(self.plugin.pp_group) self.dp_size = dist.get_world_size(self.plugin.dp_group) self.tp_size = dist.get_world_size(self.plugin.tp_group) self.pp_size = dist.get_world_size(self.plugin.pp_group) # Init Hybrid ray process group for i in range(self.num_producers): cc.init_collective_group(self.world_size + 1, self.rank + 1, group_name=f"sync_data_{i}") if self.pp_size > 1: # use hybrid tp + pp if self.tp_rank == 0 and self.dp_rank == 0: cc.init_collective_group( self.num_producers + 1, self.num_producers, group_name=f"sync_model_{self.pp_rank}" ) else: if self.rank == 0: cc.init_collective_group(self.num_producers + 1, self.num_producers, group_name="sync_model") self.buffer = [] self.recv_cnt = 0 self.profiler = CustomProfiler(f"C{self.rank}", disabled=not self.enable_profiling) def state_dict(self) -> Dict[str, torch.Tensor]: raise NotImplementedError def step(self, step_idx: int, **kwargs) -> Optional[float]: raise NotImplementedError def prepare_mini_batch(self, effective_group_to_raw_group_mapping: Dict[int, int]) -> Dict[str, torch.Tensor]: """ Prepare a mini-batch from the effective group to raw group mapping. This method is used to create a mini-batch for training. """ batches = [ self.buffer[effective_group_to_raw_group_mapping[i]] for i in range(self.dp_rank * self.minibatch_size, (self.dp_rank + 1) * self.minibatch_size) ] # every dp_rank will receive a complete mini-batch, no need to sync within step() later # each mini-batch use the first self.dp_size * minibatch_size effective samples raw_mini_batches = self.buffer[ : effective_group_to_raw_group_mapping[self.dp_size * self.minibatch_size - 1] + 1 ] # include the last effective sample raw_mini_batches_metric_dict = { "raw_train_mini_batch_reward": [t[1] for t in raw_mini_batches], "raw_train_mini_batch_format_acc": [t[2] for t in raw_mini_batches], "raw_train_mini_batch_ans_acc": [t[3] for t in raw_mini_batches], "raw_train_mini_batch_response_len": [t[4] for t in raw_mini_batches], } batch = bind_batch([t[0] for t in batches]) batch = post_recv(batch) return batch, raw_mini_batches_metric_dict def calculate_effective_group_to_raw_group_mapping(self, step): effective_group_to_raw_group_mapping = {} for buffer_idx in range(len(self.buffer)): if self.buffer[buffer_idx][0] is not None: if self.n_behind == 0: effective_group_to_raw_group_mapping[len(effective_group_to_raw_group_mapping)] = buffer_idx else: if self.buffer[buffer_idx][-1] <= step - self.n_behind: effective_group_to_raw_group_mapping[len(effective_group_to_raw_group_mapping)] = buffer_idx return effective_group_to_raw_group_mapping def loop(self) -> None: self.profiler.enter("sync_model") torch.cuda.empty_cache() state_dict = self.state_dict() if self.pp_size > 1: if self.tp_rank == 0 and self.dp_rank == 0: ray_broadcast_tensor_dict( state_dict, src=self.num_producers, device=self.device, group_name=f"sync_model_{self.pp_rank}", ) else: if self.rank == 0: ray_broadcast_tensor_dict( state_dict, src=self.num_producers, device=self.device, group_name="sync_model" ) del state_dict torch.cuda.empty_cache() self.profiler.exit("sync_model") print( f"Consumer{self.rank} num_update: {self.num_update_per_episode}, num_recv: {self.num_recv_per_update}, nmb: {self.num_microbatches}" ) for episode in range(self.num_episodes): with tqdm( range(self.num_update_per_episode), desc=f"Episode {episode} with rollout step(s)", disable=self.rank != 0, ) as pbar: for step in pbar: torch.cuda.reset_peak_memory_stats() i = 0 self.profiler.enter(f"rollout_episode_{episode}_step_{step}") for _ in range(self.num_recv_per_update): if self.n_behind > 0: # after sync model, do not wait for more data to arrive as rollout takes time, use buffered data effective_group_to_raw_group_mapping = self.calculate_effective_group_to_raw_group_mapping( step=step ) while len(effective_group_to_raw_group_mapping) >= self.dp_size * self.minibatch_size: self.profiler.log( f"Still have {len(effective_group_to_raw_group_mapping)} effective groups, greater than {self.dp_size * self.minibatch_size}, start training" ) batch, raw_mini_batches_metric_dict = self.prepare_mini_batch( effective_group_to_raw_group_mapping ) self.profiler.enter("step") loss = self.step(i, pbar, **batch, **raw_mini_batches_metric_dict) self.profiler.exit("step") self.buffer = self.buffer[ effective_group_to_raw_group_mapping[self.dp_size * self.minibatch_size - 1] + 1 : ] # recalculate the effective group to raw group mapping effective_group_to_raw_group_mapping_size_before = len( effective_group_to_raw_group_mapping ) effective_group_to_raw_group_mapping = ( self.calculate_effective_group_to_raw_group_mapping(step=step) ) assert ( len(effective_group_to_raw_group_mapping) == effective_group_to_raw_group_mapping_size_before - self.dp_size * self.minibatch_size ) if loss is not None: pbar.set_postfix({"loss": loss}) i += 1 # receive data from producers for r in range(self.num_producers): print(f"[T{dist.get_rank()}] Recv data episode {episode} step {step} from {r}") self.profiler.enter(f"recv_broadcast_data_P{r}") raw_batch = ray_broadcast_tensor_dict( None, src=0, device=self.device, group_name=f"sync_data_{r}" ) self.profiler.exit(f"recv_broadcast_data_P{r}") # calculate group reward et al. filtering. As only the filtered group will be used for training (which is incomplete), # we need to calculate the metrics before filtering here for logging # [batch_size, num_generations, ...] -> [batch_size * num_generations, ...] raw_batch = { k: v.view(-1, self.num_generations, v.size(-1)) if k != "temperature" else v for k, v in raw_batch.items() } # [batch_size, num_generations] -> [batch_size] self.total_prompt_trained += raw_batch["reward"].size(0) reward = raw_batch["reward"][:, :, 0] format_acc = raw_batch["format_acc"][:, :, 0] ans_acc = raw_batch["ans_acc"][:, :, 0] response_len = ( raw_batch["response_idx"][:, :, 1] - raw_batch["response_idx"][:, :, 0] + 1 ).type(torch.float32) effective_group_mask = None if self.filter_range is not None and self.grpo_config.get("dynamic_batching", True): # filter the group based on the reward and accuracy group_ans_acc_mean = ans_acc.mean(dim=1) effective_group_mask = torch.logical_and( group_ans_acc_mean > self.filter_range[0], group_ans_acc_mean < self.filter_range[1] ) raw_batch = unbind_batch(raw_batch) # List[Dict[str, torch.Tensor]] for group_idx, group_with_reward in enumerate(raw_batch): self.buffer.append( [ ( group_with_reward if effective_group_mask is None or effective_group_mask[group_idx] else None ), reward[group_idx], format_acc[group_idx], ans_acc[group_idx], response_len[group_idx], step, ] ) if effective_group_mask is not None: print( f"[T{dist.get_rank()}] Filter recv data: {len(raw_batch)} -> {torch.sum(effective_group_mask).cpu().item()} effective groups" ) # mapping the effective group to the raw group for indexing effective_group_to_raw_group_mapping = self.calculate_effective_group_to_raw_group_mapping( step=step ) print( f"[T{dist.get_rank()}] Collect Effective Prompt: {len(effective_group_to_raw_group_mapping)}/{self.dp_size * self.minibatch_size}" ) if self.n_behind == 0: # If n_behind is 0, we start training after receiving data from producers. while len(effective_group_to_raw_group_mapping) >= self.dp_size * self.minibatch_size: self.profiler.log( f"Collect {len(effective_group_to_raw_group_mapping)} effective groups, greater than {self.dp_size * self.minibatch_size}, start training" ) batch, raw_mini_batches_metric_dict = self.prepare_mini_batch( effective_group_to_raw_group_mapping ) self.profiler.enter("step") loss = self.step(i, pbar, **batch, **raw_mini_batches_metric_dict) self.profiler.exit("step") self.buffer = self.buffer[ effective_group_to_raw_group_mapping[self.dp_size * self.minibatch_size - 1] + 1 : ] # recalculate the effective group to raw group mapping effective_group_to_raw_group_mapping_size_before = len( effective_group_to_raw_group_mapping ) effective_group_to_raw_group_mapping = ( self.calculate_effective_group_to_raw_group_mapping(step=step) ) assert ( len(effective_group_to_raw_group_mapping) == effective_group_to_raw_group_mapping_size_before - self.dp_size * self.minibatch_size ) if loss is not None: pbar.set_postfix({"loss": loss}) i += 1 if self.lr_scheduler is not None: self.lr_scheduler.step() if (step + 1) % self.save_interval == 0 or (step + 1) == self.num_update_per_episode: if self.rank == 0: print(f"Start saving policy model at step {step + 1}.") save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.policy_model, optimizer=self.optimizer, lr_scheduler=self.lr_scheduler, epoch=episode, step=step, batch_size=int(self.total_prompt_trained / step), coordinator=self.coordinator, ) # for setting start index when resuming training if self.rank == 0: print(f"Saved model checkpoint at step {step + 1} in folder {self.save_dir}") if (episode != self.num_episodes - 1 or step != self.num_update_per_episode - 1) and ( episode != 0 or step >= self.n_behind ): if self.pp_size > 1: print( f"[T{dist.get_rank()}] Sync model PP stage {self.pp_rank} episode {episode} step {step}" ) else: print(f"[T{dist.get_rank()}] Sync model episode {episode} step {step}") self.profiler.enter("sync_model") torch.cuda.empty_cache() state_dict = self.state_dict() if self.pp_size > 1: if self.tp_rank == 0 and self.dp_rank == 0: ray_broadcast_tensor_dict( state_dict, src=self.num_producers, device=self.device, group_name=f"sync_model_{self.pp_rank}", ) else: if self.rank == 0: ray_broadcast_tensor_dict( state_dict, src=self.num_producers, device=self.device, group_name="sync_model" ) del state_dict torch.cuda.empty_cache() self.profiler.exit("sync_model") self.profiler.log(f"Peak memory usage: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") self.profiler.exit(f"rollout_episode_{episode}_step_{step}") def __del__(self): if hasattr(self, "profiler"): self.profiler.close() @ray.remote class SimpleConsumer(BaseConsumer): def __init__( self, num_producers, num_episodes, rank, world_size, master_addr, master_port, num_update_per_episode, num_recv_per_update, batch_size, model_config, plugin_config, minibatch_size=1, save_interval: int = 100, save_dir="./model", ): super().__init__( num_producers, num_episodes, rank, world_size, master_addr, master_port, num_update_per_episode, num_recv_per_update, batch_size, model_config, plugin_config, minibatch_size, save_interval, save_dir, ) path = model_config.pop("path") self.model = AutoModelForCausalLM.from_pretrained(path, **model_config) self.model.train() self.model.gradient_checkpointing_enable() self.optimizer = HybridAdam(self.model.parameters(), lr=1e-3, weight_decay=0.01) self.accum_loss = torch.zeros(1, device=self.device) def setup(self): super().setup() self.model, self.optimizer, *_ = self.booster.boost(self.model, self.optimizer) def step(self, step_idx: int, pbar: Any, **kwargs) -> Optional[float]: labels = kwargs["input_ids"].clone() labels[kwargs["attention_mask"] == 0] = -100 kwargs["labels"] = labels assert kwargs.pop("action_mask").shape == kwargs.pop("action_log_probs").shape need_update = (step_idx + 1) % self.num_microbatches == 0 ctx = nullcontext() if need_update else self.booster.no_sync(self.model, self.optimizer) with ctx: out = self.model(**kwargs) loss = out.loss / self.num_microbatches self.accum_loss.add_(loss.data) self.booster.backward(loss, self.optimizer) if need_update: self.optimizer.step() self.optimizer.zero_grad() loss_scalar = self.accum_loss.item() self.accum_loss.zero_() return loss_scalar def state_dict(self): self.model._force_wait_all_gather() model = self.model.unwrap() state_dict = model.state_dict() return state_dict
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/launch_zero_bubble.py
applications/ColossalChat/coati/distributed/launch_zero_bubble.py
import copy import os import uuid from typing import Any, Dict, Optional import ray from .comm import SharedVariableActor from .zero_bubble.distributor import Distributor from .zero_bubble.grpo_consumer import GRPOConsumer from .zero_bubble.producer import SimpleProducer ALGO_MAP = {"GRPO": GRPOConsumer, "DAPO": GRPOConsumer} def get_jsonl_size_fast(path: str) -> int: with open(path) as f: lines = f.readlines() lines = [line for line in lines if line.strip()] return len(lines) def get_dp_size_fast(n_procs: int, plugin_config: Dict[str, Any]) -> int: tp_size = plugin_config.get("tp_size", 1) pp_size = plugin_config.get("pp_size", 1) ep_size = plugin_config.get("ep_size", 1) sp_size = plugin_config.get("sp_size", 1) return n_procs // (tp_size * pp_size * ep_size * sp_size) def launch_distributed( num_producers: int, num_proc_per_producer: int, num_consumer_procs: int, num_episodes: int, inference_batch_size: int, inference_microbatch_size: int, train_batch_size: int, train_minibatch_size: int, train_dataset_config: Dict[str, Any], inference_model_config: Dict[str, Any], generate_config: Dict[str, Any], train_model_config: Dict[str, Any], grpo_config: Dict[str, Any], plugin_config: Dict[str, Any], tokenizer_config: Optional[Dict[str, Any]] = None, inference_backend: str = "transformers", num_generations: int = 8, master_addr: str = "localhost", master_port: int = 29500, core_algo: str = "GRPO", project_name: Optional[str] = None, save_interval: int = 100, save_dir: str = "./model", eval_dataset_config: Optional[Dict[str, Any]] = None, eval_interval: int = 100, eval_save_dir: Optional[str] = None, eval_generation_config: Optional[Dict[str, Any]] = None, log_rollout_interval: int = 20, rollout_save_dir: str = "./rollout", enable_profiling: bool = False, data_actor_buffer_size_limit: int = 0, ): if core_algo not in ALGO_MAP: raise NotImplementedError(f"{core_algo} is not supported yet.") else: core_consumer = ALGO_MAP.get(core_algo, GRPOConsumer) train_dp_size = get_dp_size_fast(num_consumer_procs, plugin_config) assert (inference_batch_size * num_producers) % (train_batch_size * train_dp_size) == 0 if data_actor_buffer_size_limit <= 0: # use 2 times the train_minibatch_size as the default buffer size limit data_actor_buffer_size_limit = train_minibatch_size * train_dp_size * 2 dataset_path = train_dataset_config["path"] train_dataset_size = get_jsonl_size_fast(dataset_path) global_inference_batch_size = inference_batch_size * num_producers train_dataset_size = (train_dataset_size // global_inference_batch_size) * global_inference_batch_size run_name = f"{inference_backend}_bs_{train_batch_size * train_dp_size}_temp_{generate_config['temperature']:.01f}_top_p_{generate_config['top_p']:.02f}" wandb_group_name = str(uuid.uuid4()) rollout_log_file = os.path.join( rollout_save_dir, f"{project_name.replace(' ','_')}_run_{wandb_group_name}.jsonl", ) # Attention: Ray use complex schedualing method that consider various factors including load-balancing. # when requesting resources, it is not guaranteed that the resource comes from a node with lower node it # this go against the design principle of our implementation, and we need to manually force the schedualing, # allocating the producer to nodes with lower node id and the consumer to the resouces from nodes with higher # node id. See the reference here: https://docs.ray.io/en/latest/ray-core/scheduling/index.html#nodeaffinityschedulingstrategy nodes = ray.nodes() # every producer is associated with a data worker, data worker is responsible for moving data from the producer to all consumer shared_sync_data_actor = SharedVariableActor.remote(num_consumer_procs, data_actor_buffer_size_limit) # all producer and the consumer 0 share the same model actor, model actor only provide signal for model synchronization shared_signal_actor = SharedVariableActor.remote() node_info = { node["NodeID"]: { "num_gpus": node["Resources"].get("GPU", 0), "address": node["NodeManagerAddress"], } # Default to 0 if no GPUs are available for node in nodes } gpu_to_node_id = [] gpu_to_ip_address = [] for node_id in node_info: for idx in range(int(node_info[node_id]["num_gpus"])): gpu_to_node_id.append(node_id) gpu_to_ip_address.append(node_info[node_id]["address"]) print(node_info) producer_procs = [] for i in range(num_producers): node_id = gpu_to_node_id[0] producer_ip_address = gpu_to_ip_address[0] for _ in range(num_proc_per_producer): gpu_to_node_id.pop(0) gpu_to_ip_address.pop(0) print(f"Schedual Producer P[{i}] which requires {num_proc_per_producer} GPUs on node {producer_ip_address}") producer = SimpleProducer.options(num_gpus=num_proc_per_producer, num_cpus=4).remote( shared_sync_data_actor=shared_sync_data_actor, shared_signal_actor=shared_signal_actor, producer_idx=i, num_producers=num_producers, num_consumer_procs=num_consumer_procs, num_episodes=num_episodes, batch_size=inference_batch_size, train_dataset_config=train_dataset_config, model_config=inference_model_config, generate_config=generate_config, tokenizer_config=copy.deepcopy(tokenizer_config), microbatch_size=inference_microbatch_size, backend=inference_backend, num_generations=num_generations, consumer_plugin_config=plugin_config, eval_dataset_config=eval_dataset_config, eval_interval=eval_interval, grpo_config=grpo_config, eval_save_dir=eval_save_dir, eval_generation_config=eval_generation_config, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, log_rollout_interval=log_rollout_interval, rollout_log_file=rollout_log_file, enable_profiling=enable_profiling, ) producer_procs.append(producer) # ray.get([p.setup.remote() for p in producer_procs]) generate_config_consumer = copy.deepcopy(generate_config) generate_config_consumer.update( dict( backend=inference_backend, ) ) consumer_master_ip_address = gpu_to_ip_address[0] print(f"Use {consumer_master_ip_address} as master address for torch DDP.") consumer_procs = [] for i in range(num_consumer_procs): node_id = gpu_to_node_id[0] consumer_ip_address = gpu_to_ip_address[0] gpu_to_node_id.pop(0) gpu_to_ip_address.pop(0) print(f"Schedual Consumer T[{i}] which requires 1 GPUs on node {consumer_ip_address}") consumer = core_consumer.options(num_gpus=1, num_cpus=4).remote( shared_sync_data_actor=shared_sync_data_actor, shared_signal_actor=shared_signal_actor, num_producers=num_producers, num_episodes=num_episodes, rank=i, world_size=num_consumer_procs, master_addr=consumer_master_ip_address, master_port=master_port, train_dataset_size=train_dataset_size, batch_size=train_batch_size, model_config=train_model_config, plugin_config=plugin_config, minibatch_size=train_minibatch_size, tokenizer_config=copy.deepcopy(tokenizer_config), generate_config=generate_config_consumer, grpo_config=grpo_config, num_generations=num_generations, save_interval=save_interval, save_dir=save_dir, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, enable_profiling=enable_profiling, ) consumer_procs.append(consumer) distributor_procs = [] for i in range(num_producers): distributor_procs.append( Distributor.options(num_cpus=2).remote( i, plugin_config.get("pp_size", 1), num_producers, shared_signal_actor, enable_profiling=enable_profiling, ) ) print("=================== All processes are created, starting setup torch DDP ===================", flush=True) ray.get([p.setup.remote() for p in consumer_procs]) print( "=================== All processes are setup, starting initialize communication groups ===================", flush=True, ) remote_refs = [] # Initialize consumer communication group for i, p in enumerate(consumer_procs): remote_refs.append(p.init_collective_group.remote(num_consumer_procs, i, "gloo", f"consumer_pg")) ray.get(remote_refs) remote_refs = [] # Initialize producer communication group for i, p in enumerate(producer_procs): remote_refs.append(p.init_collective_group.remote(num_producers, i, "nccl", f"producer_pg")) ray.get(remote_refs) remote_refs = [] # Initialize distributor communication group for i, p in enumerate(distributor_procs): remote_refs.append(p.init_collective_group.remote(num_producers, i, "gloo", f"distributor_pg")) ray.get(remote_refs) remote_refs = [] # Initialize sync model communication group between consumer and sync model actor # As per tested, gloo do not support nested initialization, so we need to initialize all participants in the same group in the same ray.get call. consumer_pp = plugin_config.get("pp_size", 1) for i, p in enumerate(consumer_procs): consumer_ddp_config = ray.get(p.get_ddp_config.remote()) if consumer_pp > 1: if consumer_ddp_config["tp_rank"] == 0 and consumer_ddp_config["dp_rank"] == 0: pp_rank = consumer_ddp_config["pp_rank"] remote_refs.append( p.init_collective_group.remote( num_producers + 1, 0, backend="gloo", group_name=f"sync_model_consumer_pp_{pp_rank}", gloo_timeout=3000000, ) ) for distributor_id, p_distributor in enumerate(distributor_procs): remote_refs.append( p_distributor.init_collective_group.remote( num_producers + 1, 1 + distributor_id, backend="gloo", group_name=f"sync_model_consumer_pp_{pp_rank}", gloo_timeout=3000000, ) ) ray.get(remote_refs) remote_refs = [] else: if i == 0: remote_refs.append( p.init_collective_group.remote( num_producers + 1, 0, backend="gloo", group_name=f"sync_model_consumer", gloo_timeout=3000000 ) ) for distributor_id, p_distributor in enumerate(distributor_procs): remote_refs.append( p_distributor.init_collective_group.remote( num_producers + 1, 1 + distributor_id, backend="gloo", group_name=f"sync_model_consumer", gloo_timeout=3000000, ) ) ray.get(remote_refs) remote_refs = [] # Initialize sync model communication group between producer and sync model actor for i, p in enumerate(producer_procs): if consumer_pp > 1: for pp_rank in range(consumer_pp): remote_refs.append( p.init_collective_group.remote( 2, 0, backend="gloo", group_name=f"sync_model_producer_{i}_pp_{pp_rank}", gloo_timeout=3000000 ) ) remote_refs.append( distributor_procs[i].init_collective_group.remote( 2, 1, backend="gloo", group_name=f"sync_model_producer_{i}_pp_{pp_rank}", gloo_timeout=3000000 ) ) ray.get(remote_refs) remote_refs = [] else: remote_refs.append( p.init_collective_group.remote( 2, 0, backend="gloo", group_name=f"sync_model_producer_{i}", gloo_timeout=3000000 ) ) remote_refs.append( distributor_procs[i].init_collective_group.remote( 2, 1, backend="gloo", group_name=f"sync_model_producer_{i}", gloo_timeout=3000000 ) ) ray.get(remote_refs) remote_refs = [] print("=================== All processes are set up, starting loop ===================", flush=True) ray.get([p.loop.remote() for p in (producer_procs + consumer_procs + distributor_procs)])
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/profiling_utils.py
applications/ColossalChat/coati/distributed/profiling_utils.py
import os import time class CustomProfiler: def __init__(self, name, disabled=True): self.disabled = disabled if not disabled: self.name = name self.pid = os.getpid() self.file = open(f"{name}.prof", "w") def _log(self, message): if self.disabled: return current_time = time.time() self.file.write(f"{current_time} {self.name} {self.pid}:: {message}\n") self.file.flush() def log(self, message): if self.disabled: return current_time = time.time() self.file.write(f"[Log]: {current_time} {self.name} {self.pid}:: {message}\n") self.file.flush() def enter(self, event_name): self._log(f"Enter {event_name}") def exit(self, event_name): self._log(f"Exit {event_name}") def close(self): if self.disabled: return self.file.close() print(f"Profiler data written to {self.name}.prof")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/producer.py
applications/ColossalChat/coati/distributed/producer.py
import copy import json import os from typing import Any, Dict, Optional import ray import ray.util.collective as cc import torch import tqdm import wandb from coati.dataset import StatefulDistributedSampler from coati.dataset.loader import RawConversationDataset, collate_fn_grpo from coati.distributed.profiling_utils import CustomProfiler from coati.distributed.reward.reward_fn import boxed_math_reward_fn, code_reward_fn, math_reward_fn from coati.distributed.reward.verifiable_reward import VerifiableReward from coati.utils import load_checkpoint from ray.util.collective import allreduce from ray.util.collective.types import Backend, ReduceOp from torch.utils.data import DataLoader, DistributedSampler from transformers import AutoTokenizer from colossalai.utils import get_current_device from .comm import ray_broadcast_tensor_dict from .inference_backend import BACKEND_MAP from .utils import pre_send, safe_append_to_jsonl_file try: from vllm import SamplingParams except ImportError: LLM = None class BaseProducer: def __init__( self, producer_idx: int, num_producers: int, num_consumer_procs: int, num_episodes: int, batch_size: int, train_dataset_config: Dict[str, Any], model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer_config: Optional[Dict[str, Any]] = None, microbatch_size: int = 1, backend: str = "transformers", consumer_plugin_config: Dict[str, Any] = None, eval_dataset_config=None, eval_interval=-1, # disable evaluation grpo_config: Dict[str, Any] = None, eval_save_dir: str = "./eval", project_name: str = None, run_name: str = None, wandb_group_name: str = None, log_rollout_interval: int = 20, rollout_log_file: str = "./rollout_log.jsonl", enable_profiling: bool = False, n_behind: int = 0, ): self.producer_idx = producer_idx self.num_producers = num_producers self.num_consumer_procs = num_consumer_procs self.num_episodes = num_episodes self.batch_size = batch_size self.microbatch_size = microbatch_size assert batch_size % microbatch_size == 0 self.num_microbatches = batch_size // microbatch_size self.latest_eval_step = -1 self.profiler = CustomProfiler(f"P{self.producer_idx}", disabled=not enable_profiling) self.train_dataset_config = train_dataset_config self.checkpoint_path = model_config.pop("checkpoint_path", None) self.model_config = model_config self.generate_config = generate_config self.tokenizer_config = tokenizer_config self.consumer_plugin_config = consumer_plugin_config self.eval_interval = eval_interval self.eval_save_dir = eval_save_dir self.consumer_global_step = 0 self.eval_mode = False self.log_rollout_interval = log_rollout_interval self.latest_rollout_log_step = -1 self.grpo_config = grpo_config self.n_behind = n_behind reward_model_kwargs = { k: v for k, v in grpo_config.items() if k in ["soft_over_length_punishment", "max_new_tokens", "cache_length", "code_verifier_api_url"] } self.response_format_tags = grpo_config.get("response_format_tags", None) if producer_idx == 0: if os.path.exists(rollout_log_file): raise ValueError( f"Rollout log file {rollout_log_file} already exists. Please delete it or change the name." ) else: os.makedirs(os.path.dirname(rollout_log_file), exist_ok=True) self.rollout_log_file = open(rollout_log_file, "w", encoding="utf8") if self.producer_idx == 0: self.wandb_run = wandb.init( project=project_name, sync_tensorboard=False, dir="./wandb", name=run_name + "_eval", group=wandb_group_name, ) if os.path.exists(self.eval_save_dir) and self.eval_interval > 0: raise ValueError(f"Eval save dir {self.eval_save_dir} already exists. Please delete it or change the name.") # init tokenizer if tokenizer_config is None: tokenizer_path = model_config["path"] self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path) else: tokenizer_path = tokenizer_config.pop("path") self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path, **tokenizer_config) self.tokenizer.padding_side = "left" if self.tokenizer.pad_token_id is None: self.tokenizer.pad_token = self.tokenizer.eos_token # init dataloader train_dataset_path = train_dataset_config.pop("path") self.train_dataset = RawConversationDataset(self.tokenizer, train_dataset_path, **train_dataset_config) self.train_dataloader = DataLoader( self.train_dataset, batch_size=microbatch_size, sampler=StatefulDistributedSampler( self.train_dataset, num_replicas=num_producers, rank=producer_idx, shuffle=True, drop_last=True, seed=42, ), num_workers=4, drop_last=True, collate_fn=collate_fn_grpo, ) if self.checkpoint_path is not None: # resume training from checkpoint start_epoch, start_step, sampler_start_idx = load_checkpoint(self.checkpoint_path, None, None, None, None) self.train_dataloader.sampler.set_start_index(sampler_start_idx) print( f"[P{self.producer_idx}] Resume training from checkpoint {self.checkpoint_path}, start epoch {start_epoch}, start step {start_step}, sampler start index {sampler_start_idx}" ) if grpo_config["reward_fn_type"] == "think_answer_tags": self.evaluation_function = math_reward_fn elif grpo_config["reward_fn_type"] == "boxed": self.evaluation_function = boxed_math_reward_fn elif grpo_config["reward_fn_type"] == "code": self.evaluation_function = code_reward_fn else: raise ValueError(f"Unknown evaluation function type {grpo_config['reward_fn_type']}") self.eval_dataset_config = eval_dataset_config if self.eval_dataset_config is not None: self.eval_dataloaders = {} for eval_task_name in self.eval_dataset_config: eval_dataset_path = eval_dataset_config[eval_task_name].pop("path") eval_dataset = RawConversationDataset( self.tokenizer, eval_dataset_path, **eval_dataset_config[eval_task_name] ) print(f"[P{self.producer_idx}] eval dataset {eval_task_name} size: {len(eval_dataset)}") self.eval_dataloaders[eval_task_name] = DataLoader( eval_dataset, batch_size=microbatch_size, sampler=DistributedSampler( eval_dataset, num_replicas=num_producers, rank=producer_idx, shuffle=False, drop_last=False, seed=42, ), collate_fn=collate_fn_grpo, ) else: print("No eval dataset provided, skip eval") self.device = get_current_device() self.reward_model = VerifiableReward( reward_fns=[self.evaluation_function], # multiple reward functions can be added here tokenizer=self.tokenizer, tags=self.response_format_tags, **reward_model_kwargs, ) # init backend if backend in BACKEND_MAP: self.backend_cls = BACKEND_MAP[backend] else: raise ValueError(f"Unexpected backend {backend}") self.consumer_pp_size = consumer_plugin_config.get("pp_size", 1) # consumer pp size def setup(self) -> None: cc.init_collective_group( world_size=self.num_producers, rank=self.producer_idx, backend=Backend.NCCL, group_name="producer_group", ) cc.init_collective_group(1 + self.num_consumer_procs, 0, group_name=f"sync_data_{self.producer_idx}") if self.consumer_pp_size > 1: for i in range(self.consumer_pp_size): cc.init_collective_group(self.num_producers + 1, self.producer_idx, group_name=f"sync_model_{i}") else: cc.init_collective_group(self.num_producers + 1, self.producer_idx, group_name="sync_model") def rollout(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: raise NotImplementedError def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: raise NotImplementedError def loop(self) -> None: torch.cuda.empty_cache() self.profiler.enter("sync_model") if self.consumer_pp_size > 1: for pp_idx in range(self.consumer_pp_size): state_dict = ray_broadcast_tensor_dict( None, self.num_producers, device=self.device, group_name=f"sync_model_{pp_idx}" ) if "consumer_global_step" in state_dict: self.consumer_global_step = state_dict.pop("consumer_global_step").item() self.load_state_dict(state_dict) else: state_dict = ray_broadcast_tensor_dict( None, self.num_producers, device=self.device, group_name="sync_model" ) if "consumer_global_step" in state_dict: self.consumer_global_step = state_dict.pop("consumer_global_step").item() self.load_state_dict(state_dict) self.profiler.exit("sync_model") print(f"[P{self.producer_idx}] Sync initial model done.") del state_dict torch.cuda.empty_cache() num_update_per_episode = len(self.train_dataloader) // self.num_microbatches num_valid_microbatches = num_update_per_episode * self.num_microbatches print( f"[P{self.producer_idx}] num_valid_microbatches {num_valid_microbatches}, nmb: {self.num_microbatches}, dl: {len(self.train_dataloader)}" ) for episode in range(self.num_episodes): self.train_dataloader.sampler.set_epoch(episode) for i, batch in enumerate(self.train_dataloader): if i >= num_valid_microbatches: break if self.eval_interval > 0 and self.eval_dataset_config is not None: if ( self.consumer_global_step - self.latest_eval_step >= self.eval_interval and self.consumer_global_step > self.latest_eval_step ) or self.latest_eval_step == -1: to_log_msg = {} self.eval_mode = True for eval_task_name in self.eval_dataloaders: if self.producer_idx == 0: print( f"[P{self.producer_idx}] Evaluate model at training step {self.consumer_global_step} on task {eval_task_name}" ) eval_results = [] eval_statistics_tensor = torch.zeros((2,), dtype=torch.float32).to(self.device) for eval_batch in tqdm.tqdm( self.eval_dataloaders[eval_task_name], disable=self.producer_idx != 0 ): eval_outputs = self.rollout(**eval_batch, sample_params=self.eval_sample_params) eval_results = eval_results + [ self.evaluation_function( eval_outputs["input_ids"][m][n], eval_outputs[ ( "test_cases" if self.grpo_config["reward_fn_type"] == "code" else "gt_answer" ) ][m], eval_outputs["response_idx"][m][n], tokenizer=self.tokenizer, eval_mode=True, tags=self.response_format_tags, ) for m in range(eval_outputs["input_ids"].size(0)) for n in range(eval_outputs["input_ids"].size(1)) ] eval_statistics_tensor[0] += sum([max(0, res["ans_valid"]) for res in eval_results]) eval_statistics_tensor[1] += len(eval_results) allreduce(eval_statistics_tensor, op=ReduceOp.SUM, group_name="producer_group") to_log_msg[f"eval/{eval_task_name}"] = ( eval_statistics_tensor[0].item() / eval_statistics_tensor[1].item() ) if self.producer_idx == 0: print( f"[P{self.producer_idx}]: Accuracy on {eval_task_name}: {to_log_msg[f'eval/{eval_task_name}']}" ) # save eval results safe_append_to_jsonl_file( os.path.join( self.eval_save_dir, f"{eval_task_name}_training_step_{self.consumer_global_step}.jsonl", ), eval_results, ) if self.producer_idx == 0: self.wandb_run.log(to_log_msg, step=self.consumer_global_step) self.eval_mode = False self.latest_eval_step = self.consumer_global_step self.profiler.enter("rollout") outputs = self.rollout(**batch) self.profiler.exit("rollout") outputs["temperature"] = torch.tensor( [self.model.generate_config["temperature"]] * outputs["input_ids"].size(0) ).to(outputs["input_ids"].device) bs, num_gen = outputs["input_ids"].size(0), outputs["input_ids"].size(1) self.profiler.enter("calculate_reward") if self.grpo_config["reward_fn_type"] == "code": test_cases = [] for prompt_id in range(bs): test_cases.extend([outputs["test_cases"][prompt_id]] * num_gen) reward_model_output = self.reward_model( outputs["input_ids"].view((-1, outputs["input_ids"].size(-1))), test_cases=test_cases, response_idx=outputs["response_idx"].view((-1, 2)), ) else: gt_answer = [] for prompt_id in range(bs): gt_answer.extend([outputs["gt_answer"][prompt_id]] * num_gen) reward_model_output = self.reward_model( outputs["input_ids"].view((-1, outputs["input_ids"].size(-1))), gt_answer=gt_answer, response_idx=outputs["response_idx"].view((-1, 2)), ) outputs["reward"] = ( torch.tensor([value[0] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) outputs["format_acc"] = ( torch.tensor([value[1] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) outputs["ans_acc"] = ( torch.tensor([value[2] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) if "gt_answer" in outputs: outputs.pop("gt_answer") if "test_cases" in outputs: outputs.pop("test_cases") self.profiler.exit("calculate_reward") print(f"[P{self.producer_idx}] Send data {[(k, v.shape) for k, v in outputs.items()]}") outputs = pre_send(outputs) self.profiler.enter("send_broadcast_data") ray_broadcast_tensor_dict( outputs, src=0, device=self.device, group_name=f"sync_data_{self.producer_idx}" ) self.profiler.exit("send_broadcast_data") if ( (i + 1) % self.num_microbatches == 0 and (episode != self.num_episodes - 1 or i != num_valid_microbatches - 1) and (episode != 0 or (i + 1) > self.n_behind * self.num_microbatches) ): if isinstance(self.model, BACKEND_MAP["vllm"]) and self.model.model_config.get( "enable_sleep_mode", False ): self.model.llm.sleep() # revict KV_cache to avoid OOM # don't sync model for last iteration torch.cuda.empty_cache() self.profiler.enter("sync_model") if self.consumer_pp_size > 1: for pp_idx in range(self.consumer_pp_size): print( f"[P{self.producer_idx}] Sync model PP stage {pp_idx} episode {episode} step {(i + 1) // self.num_microbatches - 1}" ) state_dict = ray_broadcast_tensor_dict( None, self.num_producers, device=self.device, group_name=f"sync_model_{pp_idx}" ) if "consumer_global_step" in state_dict: self.consumer_global_step = state_dict.pop("consumer_global_step").item() self.load_state_dict(state_dict) else: print( f"[P{self.producer_idx}] Sync model episode {episode} step {(i + 1) // self.num_microbatches - 1}" ) state_dict = ray_broadcast_tensor_dict( None, self.num_producers, device=self.device, group_name="sync_model" ) if "consumer_global_step" in state_dict: self.consumer_global_step = state_dict.pop("consumer_global_step").item() self.load_state_dict(state_dict) self.profiler.exit("sync_model") del state_dict torch.cuda.empty_cache() if isinstance(self.model, BACKEND_MAP["vllm"]) and self.model.model_config.get( "enable_sleep_mode", False ): self.model.llm.wake_up() # linear annealing for 1 episode, temperature from initial to 0.9 if episode <= 0: ratio = 1 - (len(self.train_dataloader) - i) / len(self.train_dataloader) self.model.generate_config["temperature"] = (1 - ratio) * self.generate_config[ "temperature" ] + ratio * 0.9 if isinstance(self.model, BACKEND_MAP["vllm"]): self.model.sample_params.temperature = (1 - ratio) * self.generate_config[ "temperature" ] + ratio * 0.9 def __del__(self): self.profiler.close() @ray.remote class SimpleProducer(BaseProducer): def __init__( self, producer_idx, num_producers, num_consumer_procs, num_episodes, batch_size, train_dataset_config, model_config, generate_config, tokenizer_config=None, microbatch_size=1, backend="transformers", num_generations: int = 8, consumer_plugin_config=None, eval_dataset_config=None, eval_interval=-1, # disable evaluation grpo_config: Dict[str, Any] = None, eval_save_dir: str = "./eval", eval_generation_config={}, project_name: str = None, run_name: str = None, wandb_group_name: str = None, log_rollout_interval: int = 20, rollout_log_file: str = "./rollout_log.jsonl", enable_profiling: bool = False, n_behind: int = 0, ): super().__init__( producer_idx, num_producers, num_consumer_procs, num_episodes, batch_size, train_dataset_config, model_config, generate_config, tokenizer_config, microbatch_size, backend, consumer_plugin_config, eval_dataset_config=eval_dataset_config, eval_interval=eval_interval, grpo_config=grpo_config, eval_save_dir=eval_save_dir, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, log_rollout_interval=log_rollout_interval, rollout_log_file=rollout_log_file, enable_profiling=enable_profiling, n_behind=n_behind, ) self.model = self.backend_cls(model_config, generate_config, self.tokenizer, num_generations) self.eval_generation_config = copy.deepcopy(self.model.generate_config) self.eval_generation_config["n"] = 1 # use 1 generation for evaluation self.eval_generation_config.update(eval_generation_config) self.eval_sample_params = SamplingParams(**self.eval_generation_config) @torch.no_grad() def rollout(self, input_ids, attention_mask, **kwargs): rollouts = self.model.generate(input_ids, attention_mask, **kwargs) if self.producer_idx == 0 and not self.eval_mode: if ( self.consumer_global_step - self.latest_rollout_log_step >= self.log_rollout_interval or self.latest_rollout_log_step == -1 ): new_record = ( json.dumps( { "train_step": self.consumer_global_step, "rollout": self.tokenizer.batch_decode( rollouts["input_ids"][:, 0], skip_special_tokens=True ), } ) + "\n" ) self.rollout_log_file.write(new_record) self.rollout_log_file.flush() self.latest_rollout_log_step = self.consumer_global_step return rollouts def __del__(self): if self.producer_idx == 0: self.wandb_run.finish() if hasattr(self, "rollout_log_file"): self.rollout_log_file.close() def load_state_dict(self, state_dict): self.model.load_state_dict(state_dict)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/utils.py
applications/ColossalChat/coati/distributed/utils.py
import json import os from typing import Any, Dict, List import torch from filelock import FileLock from colossalai.shardformer.layer.loss import dist_log_prob def unbind_batch(batch: Dict[str, torch.Tensor]) -> List[Dict[str, torch.Tensor]]: batches = [] for k, v in batch.items(): if len(batches) == 0: unbinded_tensors = v.unbind(0) batches = [{k: tensor} for tensor in unbinded_tensors] else: unbinded_tensors = v.unbind(0) assert len(batches) == len(unbinded_tensors) for i, tensor in enumerate(unbinded_tensors): batches[i][k] = tensor return batches def bind_batch(batches: List[Dict[str, torch.Tensor]]) -> Dict[str, torch.Tensor]: batch = {} for k in batches[0].keys(): batch[k] = torch.stack([batch[k] for batch in batches], dim=0) return batch def pre_send(batch: Dict[str, torch.Tensor]) -> Dict[str, torch.Tensor]: # compress mask to save bandwidth if "attention_mask" in batch: batch["attention_mask"] = batch["attention_mask"].to(torch.bool) if "action_mask" in batch: batch["action_mask"] = batch["action_mask"].to(torch.bool) return batch def post_recv(batch: Dict[str, torch.Tensor]) -> Dict[str, torch.Tensor]: # decompress mask if "attention_mask" in batch: batch["attention_mask"] = batch["attention_mask"].to(torch.int) if "action_mask" in batch: batch["action_mask"] = batch["action_mask"].to(torch.int) return batch def update_by_default(data: Dict[str, Any], default: Dict[str, Any]) -> Dict[str, Any]: data = data.copy() for k, v in default.items(): if k not in data: data[k] = v return data def log_probs_from_logits(logits: torch.Tensor, labels: torch.Tensor) -> torch.Tensor: """ Compute the log probabilities from logits for the given labels. Args: logits (torch.Tensor): The input logits. labels (torch.Tensor): The target labels. Returns: torch.Tensor: The log probabilities corresponding to the labels. """ log_probs = torch.log_softmax(logits, dim=-1) per_label_logps = log_probs.gather(dim=-1, index=labels.unsqueeze(-1)) return per_label_logps.squeeze(-1) def memory_efficient_logprob( logits: torch.Tensor, inputs: torch.Tensor, num_action: int, chunk_size: int = 2048, shard_config: Any = None, vocab_size: int = None, ) -> torch.Tensor: """ Calculate action log probs in a memory-efficient way by processing in chunks. Args: logits (torch.Tensor): Output tensor of Actor.forward.logits. inputs (torch.LongTensor): Input sequences. num_action (int): Number of actions. chunk_size (int, optional): Size of each chunk to process. Default is 2048. shard_config: Shard configuration for distributed computation. vocab_size (int, optional): Vocabulary size. Default is None. Returns: torch.Tensor: Action log probs. """ action_log_probs = torch.zeros((logits.size(0), num_action), device=logits.device, dtype=logits.dtype) context_length = logits.size(1) - num_action for i in range(action_log_probs.size(0)): # loop over each sample in the micro-batch for start in range(context_length, logits.size(1), chunk_size): end = min(start + chunk_size, logits.size(1)) # calculate log probs in chunks to save memory log_probs = dist_log_prob( inputs[i : i + 1, start - 1 : end], logits[i : i + 1, start - 1 : end], shard_config, vocab_size, logits.dtype, ) # [1, chunk_size, 1] log_probs = log_probs.squeeze(-1) action_log_probs[i, start - context_length : end - context_length] += log_probs[0] return action_log_probs def entropy_from_logits(logits: torch.Tensor) -> torch.Tensor: """ Calculate entropy Reference: https://github.com/volcengine/verl/blob/96b730bbed80292a439f0c0057d3920ab8b28d52/verl/utils/torch_functional.py#L145 """ p = torch.nn.functional.softmax(logits, dim=-1) entropy = torch.logsumexp(logits, dim=-1) - torch.sum(p * logits, dim=-1) return entropy def masked_mean(tensor: torch.Tensor, mask: torch.Tensor, dim: int = 1) -> torch.Tensor: """ Compute the masked mean of a tensor along a specified dimension. Args: tensor (torch.Tensor): The input tensor. mask (torch.Tensor): The mask tensor with the same shape as the input tensor. dim (int, optional): The dimension along which to compute the mean. Default is 1. Returns: torch.Tensor: The masked mean tensor. """ tensor = tensor * mask tensor = tensor.sum(dim=dim) mask_sum = mask.sum(dim=dim) mean = tensor / (mask_sum + 1e-8) return mean def masked_sum(tensor: torch.Tensor, mask: torch.Tensor, dim: int = 1) -> torch.Tensor: """ Compute the masked sum of a tensor along a specified dimension. Args: tensor (torch.Tensor): The input tensor. mask (torch.Tensor): The mask tensor with the same shape as the input tensor. dim (int, optional): The dimension along which to compute the sum. Default is 1. Returns: torch.Tensor: The masked sum tensor. """ tensor = tensor * mask return tensor.sum(dim=dim) def safe_append_to_jsonl_file(file_path, data): with FileLock(file_path + ".lock"): # Ensure file exists os.makedirs(os.path.dirname(file_path), exist_ok=True) with open(file_path, "a", encoding="utf8") as f: for entry in data: json_line = json.dumps(entry, ensure_ascii=False) f.write(json_line + "\n")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/comm.py
applications/ColossalChat/coati/distributed/comm.py
import copy from typing import Any, Dict import ray import ray.util.collective as cc import torch import torch.distributed.distributed_c10d as c10d from packaging.version import Version def ray_broadcast_object(obj: Any, src: int = 0, device=None, group_name: str = "default") -> Any: rank = cc.get_rank(group_name) if rank == src: if Version(torch.__version__) >= Version("2.3.0"): obj_tensor, size_tensor = c10d._object_to_tensor(obj, device=device, group=None) elif Version(torch.__version__) >= Version("1.13.0"): obj_tensor, size_tensor = c10d._object_to_tensor(obj, device=device) else: obj_tensor, size_tensor = c10d._object_to_tensor(obj) obj_tensor = obj_tensor.to(device) size_tensor = size_tensor.to(device) else: size_tensor = torch.empty(1, dtype=torch.int64, device=device) cc.broadcast(size_tensor, src, group_name) if rank != src: obj_tensor = torch.empty(size_tensor.item(), dtype=torch.uint8, device=device) cc.broadcast(obj_tensor, src, group_name) if rank != src: if Version(torch.__version__) >= Version("2.3.0"): obj = c10d._tensor_to_object(obj_tensor, size_tensor.item(), group=None) else: obj = c10d._tensor_to_object(obj, size_tensor.item()) return obj def ray_broadcast_tensor_dict( tensor_dict: Dict[str, torch.Tensor], src: int = 0, device=None, group_name: str = "default", backend: str = "nccl", offload_to_cpu: bool = False, pin_memory: bool = False, ) -> Dict[str, torch.Tensor]: rank = cc.get_rank(group_name) if tensor_dict is None: tensor_dict = {} if rank == src: metadata = [] for k, v in tensor_dict.items(): metadata.append((k, v.shape, v.dtype)) else: metadata = None metadata = ray_broadcast_object(metadata, src, device, group_name) for k, shape, dtype in metadata: if rank == src: if offload_to_cpu: tensor = tensor_dict[k].to(device) else: tensor = tensor_dict[k] else: tensor = tensor_dict.get(k, torch.zeros(shape, dtype=dtype, device=device, pin_memory=pin_memory)) if backend == "gloo" and dtype == torch.bfloat16: # Gloo does not support bfloat16, convert to float16 tensor = tensor.view(torch.float16) cc.broadcast(tensor, src, group_name) if backend == "gloo" and dtype == torch.bfloat16: # Convert back to bfloat16 if it was converted to float16 tensor = tensor.view(torch.bfloat16) if rank != src: if offload_to_cpu: tensor_dict[k] = tensor.cpu() else: tensor_dict[k] = tensor return tensor_dict @ray.remote class SharedVariableActor: def __init__(self, number_of_readers: int = 0, buffer_size_limit: int = 1000): self.data_queue = [] self.data_uid = 0 self.number_of_readers = number_of_readers self.queue_size = 0 self.signals = {} self.process_locks = {} self.signal_procs_meet_count = {} self.buffer_size_limit = buffer_size_limit def pickup_rollout_task(self, num_tasks: int): """ use queue size to control whether producers should generating new rollouts or wait for consumer to consumer more data. if queue size is less than threshold, it means consumer is consuming data fast enough, so producers can generate new rollouts. if queue size is greater than threshold, it means consumer is consuming data slowly, so producers should wait for consumer to consume more data. Any free producer can pick up the task to generate rollout then increase the queued_data_size to prevent other producer to pick up the task redundantly, Note it is not the real queue length as data may still be generating """ ret = False if self.queue_size < (self.buffer_size_limit / max(0.1, self.signals.get("sample_utilization", 1.0))): ret = True self.queue_size += num_tasks return ret def append_data(self, data): self.data_queue.append([self.data_uid, data, 0]) # [data_uid, data, access_count] self.data_uid += 1 return True def get_data(self, data_uid: int): # for multi-process data reading if not self.data_queue: # no data in the queue, return None return None to_pop_index = None ret = None for i, (uid, data, access_count) in enumerate(self.data_queue): if uid == data_uid: # found the data with the given uid self.data_queue[i][2] += 1 ret = copy.deepcopy(data) if self.data_queue[i][2] == self.number_of_readers: to_pop_index = i break if to_pop_index is not None: # remove the data from the queue if it has been accessed by all readers self.data_queue.pop(to_pop_index) self.queue_size -= data["input_ids"].size(0) return ret def acquire_process_lock(self, key: str): # atomic lock for process if key not in self.process_locks: self.process_locks[key] = 1 # locked return 0 if self.process_locks[key] == 0: self.process_locks[key] = 1 # lock the process return 0 else: return 1 def release_process_lock(self, key: str): # atomic unlock for process assert self.process_locks.get(key, 0) == 1, f"Releasing a process lock {key} that is not locked." self.process_locks[key] = 0 def set_signal(self, key: str, signal: str): self.signals[key] = signal def get_signal(self): return self.signals
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/loss.py
applications/ColossalChat/coati/distributed/loss.py
from typing import Optional import torch import torch.nn as nn from coati.distributed.utils import masked_mean, masked_sum class PolicyLoss(nn.Module): """ Policy Loss for PPO """ def __init__( self, clip_eps_low: float = 0.2, clip_eps_high: float = 0.2, beta: float = 0.01, loss_variation: str = "sample_level", adv: str = "GRPO", ) -> None: super().__init__() self.clip_eps_low = clip_eps_low self.clip_eps_high = clip_eps_high self.beta = beta self.loss_variation = loss_variation assert loss_variation in ["sample_level", "token_level"], f"Unsupported loss variation: {loss_variation}" self.adv = adv def forward( self, log_probs: torch.Tensor, old_log_probs: torch.Tensor, advantages: torch.Tensor, per_token_kl: torch.Tensor, action_mask: Optional[torch.Tensor] = None, loss_mask: Optional[torch.Tensor] = None, total_effective_tokens_in_batch: torch.Tensor = None, ) -> torch.Tensor: if action_mask is None: ratio = (log_probs - old_log_probs.detach()).exp() else: ratio = ((log_probs - old_log_probs.detach()) * action_mask).exp() surr1 = ratio * advantages surr2 = ratio.clamp(1 - self.clip_eps_low, 1 + self.clip_eps_high) * advantages if self.beta == 0: # skip kl term if kl coefficient is zero per_token_kl = 0.0 loss = -torch.min(surr1, surr2) + self.beta * per_token_kl if self.loss_variation == "sample_level": if action_mask is not None: loss = masked_mean(loss, action_mask) else: loss = loss.mean(dim=1) if loss_mask is not None: loss = loss * loss_mask loss = loss.mean() elif self.loss_variation == "token_level": if action_mask is not None: loss = masked_sum(loss, action_mask) else: loss = loss.sum(dim=1) if loss_mask is not None: loss = loss * loss_mask loss = loss.sum() / (total_effective_tokens_in_batch + 1e-8) else: raise ValueError(f"Unsupported loss variation: {self.loss_variation}") return loss, ratio.max()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/grpo_consumer.py
applications/ColossalChat/coati/distributed/grpo_consumer.py
from contextlib import nullcontext from typing import Any, Optional import ray import torch import wandb from coati.distributed.consumer import BaseConsumer from coati.distributed.loss import PolicyLoss from coati.distributed.utils import entropy_from_logits, memory_efficient_logprob from coati.trainer.utils import all_reduce_mean, all_reduce_sum from coati.utils import load_checkpoint from transformers import AutoModelForCausalLM, AutoTokenizer from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam @ray.remote class GRPOConsumer(BaseConsumer): def __init__( self, num_producers, num_episodes, rank, world_size, master_addr, master_port, num_update_per_episode, num_recv_per_update, batch_size, model_config, plugin_config, minibatch_size=1, num_generations=8, generate_config=None, grpo_config={}, save_interval: int = 100, save_dir="./model", project_name: str = None, run_name: str = None, wandb_group_name: str = None, enable_profiling: bool = False, n_behind: int = 0, ): print(f"Using GRPO config: {grpo_config}") if ( plugin_config.get("pp_size", 1) > 1 and "num_microbatches" not in plugin_config and "microbatch_size" not in plugin_config ): plugin_config["microbatch_size"] = max( 1, grpo_config.get("train_microbatch_size") // plugin_config.get("pp_size", 1) ) super().__init__( num_producers, num_episodes, rank, world_size, master_addr, master_port, num_update_per_episode, num_recv_per_update, batch_size, model_config, plugin_config, minibatch_size, save_interval=save_interval, save_dir=save_dir, enable_profiling=enable_profiling, n_behind=n_behind, ) path = model_config.pop("path") self.policy_model = AutoModelForCausalLM.from_pretrained(path, **model_config) self.policy_model.train() self.policy_model.gradient_checkpointing_enable() self.optimizer = HybridAdam( self.policy_model.parameters(), lr=grpo_config.get("lr", 1e-6), weight_decay=grpo_config.get("weight_decay", 0.01), ) self.accum_loss = torch.zeros(1, device=self.device) self.accum_kl = torch.zeros(1, device=self.device) self.accum_entropy = torch.zeros(1, device=self.device) self.accum_advantages = torch.zeros(1, device=self.device) self.raw_train_batch_reward = [] self.raw_train_batch_format_acc = [] self.raw_train_batch_ans_acc = [] self.raw_train_batch_response_len = [] self.accum_count = 0 self.generate_config = generate_config self.grpo_config = grpo_config self.project_name = project_name self.effective_sample_count = 0 self.effective_prompt_count = 0 self.project_name = project_name self.run_name = run_name self.wandb_group_name = wandb_group_name self.policy_loss_fn = PolicyLoss( clip_eps_low=grpo_config.get("clip_eps_low", 0.2), clip_eps_high=grpo_config.get("clip_eps_high", 0.2), beta=grpo_config.get("beta", 0.01), loss_variation=grpo_config.get("loss_variation", "sample_level"), adv=grpo_config.get("algo"), ) # Reference model is initialized from policy model. if self.policy_loss_fn.beta > 0: self.reference_model = AutoModelForCausalLM.from_pretrained(path, **model_config) self.reference_model.eval() self.tokenizer = AutoTokenizer.from_pretrained(path) self.pad_token_id = self.tokenizer.pad_token_id self.num_generations = num_generations self.filter_range = grpo_config.get("filter_range", None) if self.filter_range is not None: assert len(self.filter_range) == 2, "Filter range should have 2 values." self.filter_truncated_response = grpo_config.get("filter_truncated_response", False) if self.filter_truncated_response: self.max_length = 0 if "max_tokens" in self.generate_config: self.max_length = self.generate_config["max_tokens"] elif "max_new_tokens" in self.generate_config: self.max_length = self.generate_config["max_new_tokens"] else: raise ValueError( "either max_tokens (vllm) or max_new_tokens (transformers) must be set in generate_config." ) # Initialize verifiable reward. grpo_config.get("response_format_tags", None) self.global_step = 0 self.lr_scheduler = CosineAnnealingWarmupLR( optimizer=self.optimizer, total_steps=min(self.num_episodes, 4) * self.num_update_per_episode, warmup_steps=0, eta_min=0.1 * grpo_config.get("lr", 1e-6), ) self.adv = grpo_config.get("algo") def setup(self): super().setup() if (not self.plugin.pp_size > 1 and self.rank == 0) or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 and self.dp_rank == 0 ): self.wandb_run = wandb.init( project=self.project_name, sync_tensorboard=False, dir="./wandb", name=self.run_name, group=self.wandb_group_name, ) self.policy_model, self.optimizer, _, _, self.lr_scheduler = self.booster.boost( self.policy_model, self.optimizer, lr_scheduler=self.lr_scheduler ) if self.policy_loss_fn.beta > 0: self.reference_model, *_ = self.booster.boost(self.reference_model) if self.checkpoint_path is not None: load_checkpoint( self.checkpoint_path, self.booster, self.policy_model, self.optimizer, self.lr_scheduler, ) self.plugin.logger.set_level("ERROR") def step(self, step_idx: int, pbar: Any, **kwargs) -> Optional[float]: """ Step data from policy model: [{ "input_ids": torch.Tensor, "attention_mask": torch.Tensor, "action_mask": torch.Tensor, "action_log_probs": torch.Tensor, }, ...] Format: [minibatch_size, num_of_generation, prompt_length + response_length] --- <PAD>...<PAD><PROMPT>...<PROMPT><RESPONSE>...<RESPONSE><PAD>...<PAD>. """ # Reshape to [minibatch_size x num_of_generation, prompt_length + response_length] data = {k: v.view(-1, v.size(-1)) for k, v in kwargs.items() if "raw_train_mini_batch_" not in k} self.raw_train_batch_reward.extend(kwargs["raw_train_mini_batch_reward"]) self.raw_train_batch_format_acc.extend(kwargs["raw_train_mini_batch_format_acc"]) self.raw_train_batch_ans_acc.extend(kwargs["raw_train_mini_batch_ans_acc"]) self.raw_train_batch_response_len.extend(kwargs["raw_train_mini_batch_response_len"]) action_mask = data["action_mask"] num_action = action_mask.shape[1] old_action_log_probs = data["action_log_probs"] response_length = torch.sum(action_mask, dim=1).to(torch.float32) train_microbatch_size = self.grpo_config.get("train_microbatch_size", data["input_ids"].size(0)) reward = data["reward"].view((-1)) format_acc = data["format_acc"].view((-1)) ans_acc = data["ans_acc"].view((-1)) # [minibatch_size, num_generations] group_reward = reward.view(-1, self.num_generations) reward_mean = group_reward.mean(dim=1) # [minibatch_size x num_generations] reward_mean = reward_mean.repeat_interleave(self.num_generations, dim=0) if self.adv == "GRPO" or self.adv == "DAPO": reward_std = group_reward.std(dim=1).repeat_interleave(self.num_generations, dim=0) # [minibatch_size x num_generations] advantages = ((reward - reward_mean) / (reward_std + 1e-4)).unsqueeze(dim=-1) elif self.adv == "REINFORCE_PPB": # [minibatch_size x num_generations] advantages = ((reward - reward_mean)).unsqueeze(dim=-1) elif self.adv == "RLOO": advantages = ( reward * self.num_generations / (self.num_generations - 1) - reward_mean * self.num_generations / (self.num_generations - 1) ).unsqueeze(dim=-1) # [minibatch_size x num_of_generation] loss_mask = torch.ones(action_mask.size(0), device=action_mask.device).bool() # filter out overlength samples if self.filter_truncated_response and action_mask.size(1) == self.max_length: loss_mask = torch.logical_and( loss_mask, action_mask[:, -1] == False, ) if self.filter_range is not None and self.grpo_config.get("dynamic_batching", False) == False: # filter out samples with reward outside the range # if dynamic batching is enabled, we filter out out of range groups before training group_ans_acc_mean = ( ans_acc.view(-1, self.num_generations).mean(dim=1).repeat_interleave(self.num_generations, dim=-1) ) loss_mask = torch.logical_and( loss_mask, torch.logical_and( group_ans_acc_mean > self.filter_range[0], group_ans_acc_mean < self.filter_range[1], ), ) self.effective_prompt_count += group_reward.size(0) * self.dp_size mean_kl, mean_loss = [], [] if self.grpo_config.get("dynamic_batching", True): need_update = self.effective_prompt_count >= self.batch_size * self.dp_size else: # If dynamic batching is disabled, we need to use all samples for training. need_update = (step_idx + 1) % self.num_microbatches == 0 effective_samples = all_reduce_sum(torch.sum(loss_mask), self.plugin) effective_tokens_count = torch.sum(action_mask, dim=-1) * loss_mask total_effective_tokens_count = all_reduce_sum(torch.sum(effective_tokens_count), self.plugin) self.effective_sample_count += effective_samples.item() pbar.set_postfix( { "Global Step": self.global_step, "Gradient Accumulation on": f"{self.effective_prompt_count}/{self.batch_size * self.dp_size} effective prompts, {self.effective_sample_count}/{self.batch_size * self.dp_size * self.num_generations} effective samples", } ) # Gradient must be synchronized if zero2 is enabled. https://github.com/hpcaitech/ColossalAI/blob/44d4053fec005fe0b06b6bc755fdc962463145df/colossalai/booster/plugin/hybrid_parallel_plugin.py#L1500 ctx = ( nullcontext() if need_update or self.booster.plugin.zero_stage == 2 else self.booster.no_sync(self.policy_model, self.optimizer) ) with ctx: mini_batch_entropies = [] for forward_micro_batch_start in range(0, data["input_ids"].size(0), train_microbatch_size): input_ids_forward_micro_batch = data["input_ids"][ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] old_action_log_probs_micro_batch = old_action_log_probs[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] attention_mask_forward_micro_batch = data["attention_mask"][ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] action_mask_forward_micro_batch = action_mask[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] loss_mask_forward_micro_batch = ( loss_mask[forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size] if loss_mask is not None else None ) advantages_forward_micro_batch = advantages[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] if self.plugin.pp_size > 1: # Support training with PP. if self.policy_loss_fn.beta > 0: with torch.no_grad(): reference_model_outputs = self.booster.execute_pipeline( iter( [ { "input_ids": input_ids_forward_micro_batch, "attention_mask": attention_mask_forward_micro_batch, } ] ), self.reference_model, criterion=lambda outputs, inputs: torch.tensor( [0.0], device=action_mask.device ), # dummy criterion optimizer=None, return_loss=False, return_outputs=True, ) if self.booster.plugin.stage_manager.is_last_stage(): reference_action_log_probs = memory_efficient_logprob( reference_model_outputs["outputs"]["logits"] / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) else: # Dummy reference logprobs for data iterator. reference_action_log_probs = None else: reference_action_log_probs = None data_policy_forward = { "input_ids": input_ids_forward_micro_batch, "attention_mask": attention_mask_forward_micro_batch, "action_mask": action_mask_forward_micro_batch, "advantages": advantages_forward_micro_batch, "loss_mask": loss_mask_forward_micro_batch, "old_action_log_probs": old_action_log_probs_micro_batch, "source": self.rank, } if reference_action_log_probs is not None: data_policy_forward["reference_action_log_probs"] = reference_action_log_probs kl = [] def _criterion(outputs, inputs): action_logits = outputs.logits mini_batch_entropies.append( ( ((entropy_from_logits(action_logits[:, -num_action:]) * inputs["action_mask"]).sum(-1)) / inputs["action_mask"].sum(-1) ).detach() ) action_log_probs = memory_efficient_logprob( action_logits / self.generate_config["temperature"], inputs["input_ids"], num_action, shard_config=self.plugin.shard_config, ) if "reference_action_log_probs" in inputs: per_token_kl = ( torch.exp(inputs["reference_action_log_probs"] - action_log_probs) - (inputs["reference_action_log_probs"] - action_log_probs) - 1 ) appox_kl = torch.sum(per_token_kl * inputs["action_mask"], dim=-1) / torch.sum( inputs["action_mask"], dim=-1 ) kl.append(appox_kl.mean()) else: per_token_kl = 0.0 kl.append(torch.tensor(0.0)) inputs["advantages"].repeat_interleave(action_log_probs.size(-1), dim=-1) if self.adv == "REINFORCE_PPB": inputs["advantages"] = inputs["advantages"] - self.policy_loss_fn.beta * per_token_kl advantages_forward_micro_batch_mean = torch.sum( inputs["advantages"] * inputs["action_mask"] ) / (torch.sum(inputs["action_mask"]) + 1e-4) advantages_forward_micro_batch_std = torch.rsqrt( torch.sum( (inputs["advantages"] - advantages_forward_micro_batch_mean) ** 2 * inputs["action_mask"] ) / (torch.sum(inputs["action_mask"]) + 1e-4) + 1e-8 ) inputs["advantages"] = ( (inputs["advantages"] - advantages_forward_micro_batch_mean) * inputs["action_mask"] / (advantages_forward_micro_batch_std) ) per_token_kl = 0.0 loss, _ = self.policy_loss_fn( action_log_probs, inputs["old_action_log_probs"], inputs["advantages"], per_token_kl, inputs["action_mask"], loss_mask=inputs["loss_mask"], total_effective_tokens_in_batch=total_effective_tokens_count, ) return loss policy_model_outputs = self.booster.execute_pipeline( iter([data_policy_forward]), self.policy_model, criterion=_criterion, optimizer=self.optimizer, return_loss=True, return_outputs=False, ) loss = policy_model_outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): if len(kl) > 0: kl = all_reduce_mean(torch.mean(torch.stack(kl)).to(loss.device), self.plugin).data mean_kl.append(kl) mean_loss.append(all_reduce_mean(loss, self.plugin).data) else: policy_model_logits = self.policy_model( input_ids=input_ids_forward_micro_batch, attention_mask=attention_mask_forward_micro_batch, ).logits action_log_probs = memory_efficient_logprob( policy_model_logits / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) if self.policy_loss_fn.beta > 0: with torch.no_grad(): reference_model_logits = self.reference_model( input_ids=input_ids_forward_micro_batch, attention_mask=attention_mask_forward_micro_batch, ).logits reference_action_log_probs = memory_efficient_logprob( reference_model_logits / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) per_token_kl = ( torch.exp(reference_action_log_probs - action_log_probs) - (reference_action_log_probs - action_log_probs) - 1 ) kl = torch.sum(per_token_kl * action_mask_forward_micro_batch, dim=-1) / torch.sum( action_mask_forward_micro_batch, dim=-1 ) else: per_token_kl = 0.0 kl = None ( advantages_forward_micro_batch.repeat_interleave(action_log_probs.size(-1), dim=-1) - self.policy_loss_fn.beta * per_token_kl ) if self.adv == "REINFORCE_PPB": advantages_forward_micro_batch = ( advantages_forward_micro_batch - self.policy_loss_fn.beta * per_token_kl ) advantages_forward_micro_batch_mean = torch.sum( advantages_forward_micro_batch * action_mask_forward_micro_batch ) / (torch.sum(action_mask_forward_micro_batch) + 1e-4) advantages_forward_micro_batch_std = torch.rsqrt( torch.sum( (advantages_forward_micro_batch - advantages_forward_micro_batch_mean) ** 2 * action_mask_forward_micro_batch ) / (torch.sum(action_mask_forward_micro_batch) + 1e-4) + 1e-8 ) advantages_forward_micro_batch = ( (advantages_forward_micro_batch - advantages_forward_micro_batch_mean) * action_mask_forward_micro_batch / (advantages_forward_micro_batch_std) ) per_token_kl = 0.0 loss, _ = self.policy_loss_fn( action_log_probs, old_action_log_probs_micro_batch, advantages_forward_micro_batch, per_token_kl, action_mask_forward_micro_batch, loss_mask=loss_mask_forward_micro_batch, total_effective_tokens_in_batch=total_effective_tokens_count, ) self.booster.backward(loss, self.optimizer) loss = all_reduce_mean(loss, self.plugin) # Calculate accumulate value. if kl is not None: kl = all_reduce_mean(kl.mean(), self.plugin) mean_kl.append(kl.data) mean_loss.append(loss.data) mini_batch_entropies.append( all_reduce_mean( ( ( ( entropy_from_logits(policy_model_logits[:, -num_action:]) * action_mask_forward_micro_batch ).sum(-1) ) / action_mask_forward_micro_batch.sum(-1) ).detach(), self.plugin, ) ) if not self.plugin.pp_size > 1 or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 and self.dp_rank == 0 ): reward = all_reduce_mean(reward.mean(), self.plugin) format_acc = all_reduce_mean(format_acc.mean(), self.plugin) ans_acc = all_reduce_mean(ans_acc.mean(), self.plugin) advantages = all_reduce_mean(advantages.mean(), self.plugin) response_length = all_reduce_mean(response_length.mean(), self.plugin) entropy = all_reduce_mean(torch.cat(mini_batch_entropies, dim=0).mean(), self.plugin) self.accum_loss.add_(sum(mean_loss) / len(mean_loss)) self.accum_entropy.add_(entropy.data) if self.policy_loss_fn.beta > 0: self.accum_kl.add_(sum(mean_kl) / len(mean_kl)) self.accum_advantages.add_(advantages.data) self.accum_count += 1 if need_update: self.optimizer.step() self.optimizer.zero_grad() self.global_step += 1 # no need to run all reduce as raw_train_batch_* are not splited across dp rank sample_utilization = self.effective_sample_count / len(self.raw_train_batch_reward) / self.num_generations self.effective_prompt_count = 0 self.effective_sample_count = 0 loss_scalar = self.accum_loss.item() if not self.plugin.pp_size > 1 or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 ): if (not self.plugin.pp_size > 1 and self.rank == 0) or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 ): raw_batch_reward_mean = torch.cat(self.raw_train_batch_reward, dim=0).mean().cpu().item() raw_batch_format_acc_mean = torch.cat(self.raw_train_batch_format_acc, dim=0).mean().cpu().item() raw_batch_ans_acc_mean = torch.cat(self.raw_train_batch_ans_acc, dim=0).mean().cpu().item() raw_batch_response_len = torch.cat(self.raw_train_batch_response_len, dim=0) raw_batch_response_len_mean = raw_batch_response_len.mean().cpu().item() overlength_samples_ratio = ( (raw_batch_response_len >= action_mask.size(-1)).to(float).mean().cpu().item() ) # not an exact figure, but a close estimate self.raw_train_batch_reward = [] self.raw_train_batch_format_acc = [] self.raw_train_batch_ans_acc = [] self.raw_train_batch_response_len = [] to_log_msg = [ f"Loss: {self.accum_loss.item() / self.accum_count:.4f}", f"Reward: {raw_batch_reward_mean:.4f}", f"format Reward: {raw_batch_format_acc_mean:.4f}", f"Acc Reward: {raw_batch_ans_acc_mean:.4f}", f"Advantages: {self.accum_advantages.item() / self.accum_count:.4f}", f"Response Length: {raw_batch_response_len_mean:.4f}", f"Sample_utilization: {sample_utilization:.4f}", f"Overlength samples ratio: {overlength_samples_ratio:.4f}", f"Entropy: {self.accum_entropy.item() / self.accum_count:.4f}", ] + ([f"KL: {self.accum_kl.item() / self.accum_count:.4f}"] if self.policy_loss_fn.beta > 0 else []) print("\n".join(to_log_msg)) metrics = { "metrics/reward": raw_batch_reward_mean, "metrics/format_acc": raw_batch_format_acc_mean, "metrics/ans_acc": raw_batch_ans_acc_mean, "metrics/response_length": raw_batch_response_len_mean, "train/loss": self.accum_loss.item() / self.accum_count, "train/advantages": self.accum_advantages.item() / self.accum_count, "train/learning_rate": self.lr_scheduler.get_last_lr()[0], "train/sample_utilization": sample_utilization, "train/entropy": self.accum_entropy.item() / self.accum_count, "train/overlength_samples_ratio": overlength_samples_ratio, "rollout/temperature": data["temperature"].cpu().numpy()[0][0], } if self.policy_loss_fn.beta > 0: metrics["train/kl"] = self.accum_kl.item() / self.accum_count if self.wandb_run is not None: self.wandb_run.log(metrics) self.accum_loss.zero_() self.accum_kl.zero_() self.accum_entropy.zero_() self.accum_advantages.zero_() self.accum_count = 0 return loss_scalar else: return None def state_dict(self): self.policy_model._force_wait_all_gather() model = self.policy_model.unwrap() state_dict = model.state_dict() state_dict["consumer_global_step"] = torch.tensor([self.global_step], device=self.device) return state_dict
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/__init__.py
applications/ColossalChat/coati/distributed/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/launch.py
applications/ColossalChat/coati/distributed/launch.py
import copy import os import uuid from typing import Any, Dict, Optional import ray from .consumer import SimpleConsumer from .grpo_consumer import GRPOConsumer from .producer import SimpleProducer ALGO_MAP = { "Simple": SimpleConsumer, "GRPO": GRPOConsumer, "DAPO": GRPOConsumer, "REINFORCE_PPB": GRPOConsumer, "RLOO": GRPOConsumer, } def get_jsonl_size_fast(path: str) -> int: with open(path) as f: lines = f.readlines() lines = [line for line in lines if line.strip()] return len(lines) def get_dp_size_fast(n_procs: int, plugin_config: Dict[str, Any]) -> int: tp_size = plugin_config.get("tp_size", 1) pp_size = plugin_config.get("pp_size", 1) ep_size = plugin_config.get("ep_size", 1) sp_size = plugin_config.get("sp_size", 1) return n_procs // (tp_size * pp_size * ep_size * sp_size) def launch_distributed( num_producers: int, num_proc_per_producer: int, num_consumer_procs: int, num_episodes: int, inference_batch_size: int, inference_microbatch_size: int, train_batch_size: int, train_minibatch_size: int, train_dataset_config: Dict[str, Any], inference_model_config: Dict[str, Any], generate_config: Dict[str, Any], train_model_config: Dict[str, Any], grpo_config: Dict[str, Any], plugin_config: Dict[str, Any], tokenizer_config: Optional[Dict[str, Any]] = None, inference_backend: str = "transformers", num_generations: int = 8, master_addr: str = "localhost", master_port: int = 29500, core_algo: str = "GRPO", project_name: Optional[str] = None, save_interval: int = 100, save_dir: str = "./model", eval_dataset_config: Optional[Dict[str, Any]] = None, eval_interval: int = 100, eval_save_dir: Optional[str] = None, eval_generation_config: Optional[Dict[str, Any]] = None, log_rollout_interval: int = 20, rollout_save_dir: str = "./rollout", enable_profiling: bool = False, n_behind: int = 0, ): if core_algo not in ALGO_MAP: raise NotImplementedError(f"{core_algo} is not supported yet.") else: core_consumer = ALGO_MAP.get(core_algo, SimpleConsumer) train_dp_size = get_dp_size_fast(num_consumer_procs, plugin_config) assert (inference_batch_size * num_producers) % (train_batch_size * train_dp_size) == 0 dataset_path = train_dataset_config["path"] num_samples = get_jsonl_size_fast(dataset_path) global_inference_batch_size = inference_batch_size * num_producers num_update_per_episode = num_samples // global_inference_batch_size num_recv_per_update = inference_batch_size // inference_microbatch_size run_name = f"{inference_backend}_bs_{train_batch_size * train_dp_size}_temp_{generate_config['temperature']:.01f}_top_p_{generate_config['top_p']:.02f}" wandb_group_name = str(uuid.uuid4()) rollout_log_file = os.path.join( rollout_save_dir, f"{project_name.replace(' ','_')}_run_{wandb_group_name}.jsonl", ) # Attention: Ray use complex schedualing method that consider various factors including load-balancing. # when requesting resources, it is not guaranteed that the resource comes from a node with lower node it # this go against the design principle of our implementation, and we need to manually force the schedualing, # allocating the producer to nodes with lower node id and the consumer to the resouces from nodes with higher # node id. See the reference here: https://docs.ray.io/en/latest/ray-core/scheduling/index.html#nodeaffinityschedulingstrategy nodes = ray.nodes() node_info = { node["NodeID"]: { "num_gpus": node["Resources"].get("GPU", 0), "address": node["NodeManagerAddress"], } # Default to 0 if no GPUs are available for node in nodes } gpu_to_node_id = [] gpu_to_ip_address = [] for node_id in node_info: for idx in range(int(node_info[node_id]["num_gpus"])): gpu_to_node_id.append(node_id) gpu_to_ip_address.append(node_info[node_id]["address"]) print(node_info) producer_procs = [] for i in range(num_producers): node_id = gpu_to_node_id[0] producer_ip_address = gpu_to_ip_address[0] for _ in range(num_proc_per_producer): gpu_to_node_id.pop(0) gpu_to_ip_address.pop(0) print(f"Schedual Producer P[{i}] which requires {num_proc_per_producer} GPUs on node {producer_ip_address}") producer = SimpleProducer.options(num_gpus=num_proc_per_producer).remote( producer_idx=i, num_producers=num_producers, num_consumer_procs=num_consumer_procs, num_episodes=num_episodes, batch_size=inference_batch_size, train_dataset_config=train_dataset_config, model_config=inference_model_config, generate_config=generate_config, tokenizer_config=tokenizer_config, microbatch_size=inference_microbatch_size, backend=inference_backend, num_generations=num_generations, consumer_plugin_config=plugin_config, eval_dataset_config=eval_dataset_config, eval_interval=eval_interval, grpo_config=grpo_config, eval_save_dir=eval_save_dir, eval_generation_config=eval_generation_config, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, log_rollout_interval=log_rollout_interval, rollout_log_file=rollout_log_file, enable_profiling=enable_profiling, n_behind=n_behind, ) producer_procs.append(producer) ray.get([p.setup.remote() for p in producer_procs]) generate_config_consumer = copy.deepcopy(generate_config) generate_config_consumer.update( dict( backend=inference_backend, ) ) consumer_master_ip_address = gpu_to_ip_address[0] print(f"Use {consumer_master_ip_address} as master address for torch DDP.") consumer_procs = [] for i in range(num_consumer_procs): node_id = gpu_to_node_id[0] consumer_ip_address = gpu_to_ip_address[0] gpu_to_node_id.pop(0) gpu_to_ip_address.pop(0) print(f"Schedual Consumer T[{i}] which requires 1 GPUs on node {consumer_ip_address}") consumer = core_consumer.options(num_gpus=1).remote( num_producers=num_producers, num_episodes=num_episodes, rank=i, world_size=num_consumer_procs, master_addr=consumer_master_ip_address, master_port=master_port, num_update_per_episode=num_update_per_episode, num_recv_per_update=num_recv_per_update, batch_size=train_batch_size, model_config=train_model_config, plugin_config=plugin_config, minibatch_size=train_minibatch_size, generate_config=generate_config_consumer, grpo_config=grpo_config, num_generations=num_generations, save_interval=save_interval, save_dir=save_dir, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, enable_profiling=enable_profiling, n_behind=n_behind, ) consumer_procs.append(consumer) ray.get([p.setup.remote() for p in consumer_procs]) ray.get([p.loop.remote() for p in (producer_procs + consumer_procs)])
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/inference_backend.py
applications/ColossalChat/coati/distributed/inference_backend.py
from typing import Any, Dict import torch import torch.nn.functional as F from transformers import AutoConfig, AutoModelForCausalLM, PreTrainedModel, PreTrainedTokenizer from colossalai.utils import get_current_device from .utils import log_probs_from_logits, update_by_default try: import sglang as sgl except ImportError: sgl = None try: from vllm import LLM, SamplingParams except ImportError: LLM = None class BaseInferenceBackend: def __init__(self, model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer: PreTrainedTokenizer): pass def generate(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: """Generate new tokens given input_ids and attention_mask. Args: input_ids (torch.Tensor): shape [B, S] attention_mask (torch.Tensor): shape [B, S] Returns: Dict[str, torch.Tensor]: containing the - input_ids (torch.Tensor): shape [B, S+N] - attention_mask (torch.Tensor): shape [B, S+N] - action_log_probs (torch.Tensor): shape [B, N] - action_mask (torch.Tensor): shape [B, N] where N is the number of generated tokens. And all tensors should be on CUDA. """ def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: pass class TransformersInferenceBackend(BaseInferenceBackend): DEFAULT_MODEL_CONFIG = dict( trust_remote_code=True, torch_dtype=torch.bfloat16, ) FORCE_MODEL_CONFIG = dict( device_map="auto", ) FORCE_GENERATE_CONFIG = dict(output_logits=True, return_dict_in_generate=True) def __init__( self, model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer: PreTrainedTokenizer, num_generations: int = 8, tokenizer_config: Dict[str, Any] = None, ): model_config = update_by_default(model_config, self.DEFAULT_MODEL_CONFIG) model_config.update(self.FORCE_MODEL_CONFIG) path = model_config.pop("path") self.model: PreTrainedModel = AutoModelForCausalLM.from_pretrained(path, **model_config) self.generate_config = generate_config.copy() self.generate_config.update(self.FORCE_GENERATE_CONFIG) self.tokenizer = tokenizer self.num_generations = num_generations @torch.no_grad() def generate(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: micro_batch_size = input_ids.size(0) input_ids = input_ids.to(get_current_device()) attention_mask = attention_mask.to(get_current_device()) gt_answer = kwargs.pop("gt_answer", None) test_cases = kwargs.pop("test_cases", None) if self.num_generations > 1: input_ids = input_ids.repeat_interleave(self.num_generations, dim=0) attention_mask = attention_mask.repeat_interleave(self.num_generations, dim=0) out = self.model.generate( input_ids, attention_mask=attention_mask, **kwargs, **self.generate_config, tokenizer=self.tokenizer ) input_len = input_ids.shape[-1] new_token_ids = out.sequences[:, input_len:] # get log probs assert new_token_ids.shape[-1] == len(out.logits) action_log_probs = [] for i, logits in enumerate(out.logits): action_log_probs.append(log_probs_from_logits(logits[:, None, :], new_token_ids[:, i : i + 1])) action_log_probs = torch.cat(action_log_probs, dim=1) # get action mask response_idx = torch.zeros((new_token_ids.size(0), 2), dtype=torch.int).to(get_current_device()) action_mask = torch.ones_like(new_token_ids, dtype=attention_mask.dtype) if self.tokenizer.eos_token_id is not None: for indices in torch.nonzero(new_token_ids == self.tokenizer.eos_token_id): action_mask[indices[0], indices[1] + 1 :] = 0 response_idx[:, 0] = input_len response_idx[:, 1] = input_len + action_mask.sum(dim=1) - 1 if attention_mask.size(0) != action_mask.size(0): assert action_mask.size(0) % attention_mask.size(0) == 0 attention_mask = attention_mask.repeat_interleave(action_mask.size(0) // attention_mask.size(0), dim=0) attention_mask = torch.cat((attention_mask, action_mask), dim=1) data = { "input_ids": out.sequences, "attention_mask": attention_mask, "action_log_probs": action_log_probs, "action_mask": action_mask, "response_idx": response_idx, } data = {k: v.view(micro_batch_size, self.num_generations, v.size(-1)) for k, v in data.items()} if gt_answer is not None: data["gt_answer"] = gt_answer if test_cases is not None: data["test_cases"] = test_cases data = {k: v.to(get_current_device()) for k, v in data.items()} return data def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: self.model.load_state_dict(state_dict) class SGLangInferenceBackend(BaseInferenceBackend): def __init__( self, model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer: PreTrainedTokenizer, num_generations: int = 8, tokenizer_config: Dict[str, Any] = None, ): if sgl is None: raise ImportError("sglang is not installed") path = model_config.pop("path") defaut_config = dict( trust_remote_code=True, skip_tokenizer_init=True, ) defaut_config.update(model_config) self.llm = sgl.Engine(model_path=path, **defaut_config) self.generate_config = generate_config self.tokenizer = tokenizer self.config = AutoConfig.from_pretrained(path) @torch.no_grad() def generate(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: outputs = self.llm.generate(input_ids=input_ids.tolist(), sampling_params=self.generate_config) out_tokens = [] out_len = [] for out in outputs: out_tokens.append(out["token_ids"]) out_len.append(out["meta_info"]["completion_tokens"]) max_len = max(out_len) input_len = input_ids.shape[-1] attention_mask = F.pad(attention_mask, (0, max_len), value=1) for i in range(len(out_tokens)): out_tokens[i] = out_tokens[i] + [self.tokenizer.pad_token_id] * (max_len - out_len[i]) attention_mask[i, input_len + out_len[i] :] = 0 out = torch.tensor(out_tokens) out = torch.cat((input_ids, out), dim=1) labels = out.clone() labels[..., :input_len] = -100 for i in range(len(out_len)): labels[i, input_len + out_len[i] :] = -100 data = { "input_ids": out, "attention_mask": attention_mask, "labels": labels, } data = {k: v.to(get_current_device()) for k, v in data.items()} return data def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: if self.config.tie_word_embeddings: del state_dict["lm_head.weight"] named_tensors = [(k, v) for k, v in state_dict.items()] self.llm.update_weights_from_tensor(named_tensors) class VLLMInferenceBackend(BaseInferenceBackend): DEFAULT_MODEL_CONFIG = dict( trust_remote_code=True, enable_sleep_mode=False, ) FORCE_GENERATE_CONFIG = dict( logprobs=0, ) def __init__( self, model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer: PreTrainedTokenizer, num_generations: int = 8, tokenizer_config: Dict[str, Any] = None, ): if LLM is None: raise ImportError("vllm is not installed") model_config = update_by_default(model_config, self.DEFAULT_MODEL_CONFIG) path = model_config.pop("path") tokenizer_path = tokenizer_config.get("path", None) if tokenizer_config is not None else None self.llm = LLM(model=path, tokenizer=tokenizer_path, **model_config) generate_config = generate_config.copy() generate_config.update(self.FORCE_GENERATE_CONFIG) generate_config.update({"n": num_generations}) self.generate_config = generate_config self.sample_params = SamplingParams(**generate_config) self.model_config = model_config self.tokenizer = tokenizer self.num_generations = num_generations @torch.no_grad() def generate(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: micro_batch_size = input_ids.size(0) response_start_idx = input_ids.size(1) first_non_padding_token_idx = (input_ids != self.tokenizer.pad_token_id).int().argmax(dim=1) micro_batch_input_ids = input_ids.tolist() micro_batch_input_ids_no_padding = [ micro_batch_input_ids[i][first_non_padding_token_idx[i] :] for i in range(micro_batch_size) ] sample_params = kwargs.get("sample_params", self.sample_params) outputs = self.llm.generate( prompt_token_ids=micro_batch_input_ids_no_padding, sampling_params=sample_params, use_tqdm=False ) out_tokens = [] out_len = [] log_probs = [] response_idx = [] for out in outputs: for output_i in out.outputs: out_len.append(len(output_i.token_ids)) out_tokens.append(list(output_i.token_ids)) response_idx.append((response_start_idx, response_start_idx + len(output_i.token_ids) - 1)) assert len(output_i.logprobs) == len(output_i.token_ids) p = [m[t].logprob for m, t in zip(output_i.logprobs, output_i.token_ids)] log_probs.append(p) # pad them max_len = self.sample_params.max_tokens action_mask = torch.ones(len(out_tokens), max_len, dtype=attention_mask.dtype) for i, new_token_ids in enumerate(out_tokens): pad_len = max_len - out_len[i] out_tokens[i] = new_token_ids + [self.tokenizer.pad_token_id] * pad_len log_probs[i] = log_probs[i] + [0.0] * pad_len action_mask[i, out_len[i] :] = 0 out_tokens = torch.tensor(out_tokens) log_probs = torch.tensor(log_probs) response_idx = torch.tensor(response_idx) if attention_mask.size(0) != action_mask.size(0): assert action_mask.size(0) % attention_mask.size(0) == 0 num_returns = action_mask.size(0) // attention_mask.size(0) attention_mask = attention_mask.repeat_interleave(num_returns, dim=0) input_ids = input_ids.repeat_interleave(num_returns, dim=0) out_tokens = torch.cat((input_ids, out_tokens), dim=1) attention_mask = torch.cat((attention_mask, action_mask), dim=1) data = { "input_ids": out_tokens, "attention_mask": attention_mask, "action_log_probs": log_probs, "action_mask": action_mask, "response_idx": response_idx, } data = {k: v.view(micro_batch_size, -1, v.size(-1)) for k, v in data.items()} data = {k: v.to(get_current_device()) for k, v in data.items()} if "gt_answer" in kwargs: data["gt_answer"] = kwargs["gt_answer"] if "test_cases" in kwargs: data["test_cases"] = kwargs["test_cases"] return data def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: self.llm.llm_engine.model_executor.driver_worker.model_runner.model.load_weights(state_dict.items()) BACKEND_MAP = { "transformers": TransformersInferenceBackend, # "sglang": SGLangInferenceBackend, # sglang backend will stuck the process due to unknown reason "vllm": VLLMInferenceBackend, }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/reward/reward_utils.py
applications/ColossalChat/coati/distributed/reward/reward_utils.py
# Copyright Unakar # Modified from https://github.com/Unakar/Logic-RL/blob/086373176ac198c97277ff50f4b6e7e1bfe669d3/verl/utils/reward_score/kk.py#L99 # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from typing import Dict, Optional, Tuple def validate_response_structure(processed_str: str, tags: Dict = None) -> bool: """Performs comprehensive validation of response structure. Args: processed_str: Processed response string from the model Returns: Boolean indicating whether all formatting requirements are met """ validation_passed = True # Check required tags if tags is None: tags = { "think_start": {"text": "<think>", "num_occur": 1}, "think_end": {"text": "</think>", "num_occur": 1}, "answer_start": {"text": "<answer>", "num_occur": 1}, "answer_end": {"text": "</answer>", "num_occur": 1}, } positions = {} for tag_name, tag_info in tags.items(): tag_str = tag_info["text"] expected_count = tag_info["num_occur"] count = processed_str.count(tag_str) positions[tag_name] = pos = processed_str.find(tag_str) if count != expected_count: validation_passed = False # Verify tag order if ( positions["think_start"] > positions["think_end"] or positions["think_end"] > positions["answer_start"] or positions["answer_start"] > positions["answer_end"] ): validation_passed = False if len(processed_str) - positions["answer_end"] != len(tags["answer_end"]["text"]): validation_passed = False return validation_passed def extract_solution(solution_str: str) -> Tuple[Optional[str], str]: """Extracts the final answer from the model's response string. Args: solution_str: Raw response string from the language model Returns: Tuple containing (extracted_answer, processed_string) """ # Extract final answer using XML-style tags answer_pattern = r"<answer>(.*?)</answer>" matches = list(re.finditer(answer_pattern, solution_str, re.DOTALL)) if not matches: return None, solution_str final_answer = matches[-1].group(1).strip() return final_answer, solution_str def extract_boxed_solution(text: str) -> Optional[str]: """ Modified from: https://gist.github.com/lewtun/9c2ce1937b741404090a3dc4c7c022b3 Retrieves the content from the last occurrence of `\boxed{}` in a LaTeX-like string. Args: text (str): A string potentially containing LaTeX-style boxed expressions. Returns: Optional[str]: The text inside the final `\boxed{}` if successfully extracted; returns `None` if no properly closed box is found. Examples: >>> extract_boxed_solution("The answer is \\boxed{42}.") '42' >>> extract_boxed_solution("Here is an unmatched \\boxed{42") None """ try: # Find the last occurrence of "\boxed{" start_idx = text.rindex("\\boxed{") # Move past "\boxed{" to find the start of the content content_start = start_idx + len("\\boxed{") open_braces = 1 pos = content_start # Traverse the string to find the matching closing brace while open_braces > 0 and pos < len(text): if text[pos] == "{": open_braces += 1 elif text[pos] == "}": open_braces -= 1 pos += 1 # If all braces are matched, extract and return the content if open_braces == 0: return text[content_start : pos - 1].strip() else: return None except ValueError: # "\boxed{" not found return None except Exception: # Any other unexpected error return None
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/reward/reward_fn.py
applications/ColossalChat/coati/distributed/reward/reward_fn.py
# Copyright 2024 ByteDance Group # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Some functions in this file are adapted from the verl project under the Apache License 2.0: https://github.com/volcengine/verl """ import json import torch from latex2sympy2_extended import NormalizationConfig from math_verify import ExprExtractionConfig, LatexExtractionConfig, parse, verify from .code_reward.utils import check_correctness_code_api as check_correctness_code from .reward_utils import extract_boxed_solution, extract_solution, validate_response_structure CANNOT_PARSE_GT_ANSWER = -1 CANNOT_PARSE_PREDICTION = -2 SUCCESS = 1 MATCHING_FAIL = 0 def verify_math_representation(completion, gt_answer): """ Verify if the completion is a valid math representation of the gt_answer. """ if not completion.startswith("\\boxed{"): completion = "\\boxed{" + completion + "}" if not gt_answer.startswith("\\boxed{"): gt_answer = "\\boxed{" + gt_answer + "}" target = ( ExprExtractionConfig(), LatexExtractionConfig( normalization_config=NormalizationConfig( nits=False, malformed_operators=False, basic_latex=True, boxed="all", units=True, ), boxed_match_priority=0, ), ) if not isinstance(gt_answer, str) or len(gt_answer) == 0: raise ValueError("gt_answer should be a string, please verify your training data.") if not isinstance(completion, str) or len(completion) == 0: return MATCHING_FAIL try: parsed_gt_answer = parse(gt_answer, extraction_config=target) if len(parsed_gt_answer) == 0: return CANNOT_PARSE_GT_ANSWER parsed_completion = parse(completion, extraction_config=target) if len(parsed_completion) == 0: return CANNOT_PARSE_PREDICTION if verify(parsed_gt_answer, parsed_completion): return SUCCESS else: return MATCHING_FAIL except Exception: return MATCHING_FAIL def verify_model_answer(decoded_final_answer, gt_answer, ans_acc, acc_score, reward): math_verify_result = verify_math_representation(decoded_final_answer, gt_answer) exact_match_result = ( SUCCESS if decoded_final_answer.strip().replace(" ", "").replace("{", "").replace("}", "").replace(",", "") == gt_answer.strip().replace(" ", "").replace("{", "").replace("}", "").replace(",", "") else MATCHING_FAIL ) if math_verify_result == SUCCESS: ans_acc += 1 reward += acc_score elif exact_match_result == SUCCESS: # sometimes for answers that's not a (valid) math expression, math_verify will fail ans_acc += 1 if math_verify_result == CANNOT_PARSE_PREDICTION: reward += ( acc_score / 2 ) # not a valid latex math representation, but the answer is correct, receive half of the score else: reward += acc_score return reward, ans_acc def math_reward_fn(input_ids, gt_answer, response_idx, **kwargs): tokenizer = kwargs["tokenizer"] eval_mode = kwargs.get("eval_mode", False) soft_over_length_punishment = kwargs.get("soft_over_length_punishment", False) acc_score = 10.0 reward = torch.tensor(0.0) format_acc = torch.tensor(0.0) ans_acc = torch.tensor(0.0) s, e = response_idx[0], response_idx[1] length_reward = 0.0 res_length = e.item() - s.item() + 1 if not eval_mode: max_new_tokens = kwargs["max_new_tokens"] else: max_new_tokens = -1 # for eval mode, we don't need to check the length if not eval_mode and soft_over_length_punishment: cache_length = kwargs["cache_length"] if max_new_tokens - cache_length < res_length < max_new_tokens: length_reward = ((max_new_tokens - cache_length) - res_length) / cache_length * acc_score if gt_answer is None: raise ValueError("no gt_answer is provided, please check your training dataset.") decoded_final_answer = tokenizer.decode(input_ids[s : e + 1], skip_special_tokens=True) final_answer, processed_str = extract_solution(decoded_final_answer) format_valid = validate_response_structure(processed_str, kwargs["tags"]) # Check answer accuracy, answer is considered correct if the answer is correct and the format is valid if final_answer is not None: if eval_mode or format_valid: reward, ans_acc = verify_model_answer(final_answer, gt_answer, ans_acc, acc_score, reward) if not eval_mode: reward = reward + length_reward # Check format accuracy if format_valid: format_acc += 1 # Check if the sequence is over length if not eval_mode and res_length >= max_new_tokens: reward *= 0.0 if not eval_mode: return torch.tensor([reward, format_acc, ans_acc]).to(input_ids.device) else: prompt = tokenizer.decode(input_ids[:s], skip_special_tokens=True) return { "prompt": prompt, "prediction": decoded_final_answer, "gold": gt_answer, "parsed": final_answer, "format_valid": format_acc.item(), "ans_valid": ans_acc.item(), "response_length": res_length, "reward": reward.item(), } def boxed_math_reward_fn(input_ids, gt_answer, response_idx, **kwargs): tokenizer = kwargs["tokenizer"] eval_mode = kwargs.get("eval_mode", False) soft_over_length_punishment = kwargs.get("soft_over_length_punishment", False) acc_score = 10.0 reward = torch.tensor(0.0) format_acc = torch.tensor(0.0) ans_acc = torch.tensor(0.0) s, e = response_idx[0], response_idx[1] length_reward = 0.0 res_length = e.item() - s.item() + 1 if not eval_mode: max_new_tokens = kwargs["max_new_tokens"] else: max_new_tokens = -1 # for eval mode, we don't need to check the length if not eval_mode and soft_over_length_punishment: cache_length = kwargs["cache_length"] if max_new_tokens - cache_length < res_length < max_new_tokens: length_reward = ((max_new_tokens - cache_length) - res_length) / cache_length * acc_score if gt_answer is None: raise ValueError("no gt_answer is provided, please check your training dataset.") decoded_final_answer = tokenizer.decode(input_ids[s : e + 1], skip_special_tokens=True) final_answer = extract_boxed_solution(decoded_final_answer) format_valid = final_answer is not None if "tags" in kwargs and kwargs["tags"]: tags = kwargs["tags"] format_valid = format_valid and all( [decoded_final_answer.count(tags[tag]["text"]) == tags[tag]["num_occur"] for tag in tags] ) # Check answer accuracy, answer is considered correct if the answer is correct and the format is valid if final_answer is not None: if eval_mode or format_valid: reward, ans_acc = verify_model_answer(final_answer, gt_answer, ans_acc, acc_score, reward) if not eval_mode: reward = reward + length_reward # Check format accuracy if format_valid: format_acc += 1 # Check if the sequence is over length if not eval_mode and res_length >= max_new_tokens: reward *= 0.0 if not eval_mode: return torch.tensor([reward, format_acc, ans_acc]).to(input_ids.device) else: prompt = tokenizer.decode(input_ids[:s], skip_special_tokens=True) return { "prompt": prompt, "prediction": decoded_final_answer, "gold": gt_answer, "parsed": final_answer, "format_valid": format_acc.item(), "ans_valid": ans_acc.item(), "response_length": res_length, "reward": reward.item(), } def code_reward_fn(input_ids, test_cases, response_idx, **kwargs): url = kwargs.get("url", "http://localhost:8000/check_correctness") tokenizer = kwargs["tokenizer"] eval_mode = kwargs.get("eval_mode", False) soft_over_length_punishment = kwargs.get("soft_over_length_punishment", False) acc_score = 10.0 reward = torch.tensor(0.0) format_acc = torch.tensor(0.0) ans_acc = torch.tensor(0.0) s, e = response_idx[0], response_idx[1] length_reward = 0.0 res_length = e.item() - s.item() + 1 if not eval_mode: max_new_tokens = kwargs["max_new_tokens"] else: max_new_tokens = -1 # for eval mode, we don't need to check the length if not eval_mode and soft_over_length_punishment: cache_length = kwargs["cache_length"] if max_new_tokens - cache_length < res_length < max_new_tokens: length_reward = ((max_new_tokens - cache_length) - res_length) / cache_length * acc_score # try to get code solution from completion. if the completion is pure code, this will not take effect. decoded_final_answer = tokenizer.decode(input_ids[s : e + 1], skip_special_tokens=True) solution = decoded_final_answer.split("```python")[-1].split("```")[0] format_valid = False if "```python" in decoded_final_answer: format_valid = solution is not None # Check format accuracy if format_valid: format_acc += 1 res = [] metadata = [] try: try: if not isinstance(test_cases, dict): test_cases = json.loads(test_cases) except Exception as e: print(f"Error {e}: Cannot parse test cases.") raise e # Complete check on all in-out pairs first. If there is no failure, per-sample test can be skipped. try: res, metadata = check_correctness_code( in_outs=test_cases, generation=solution, timeout=10, debug=False, url=url ) metadata = dict(enumerate(metadata))[0] success = all(map(lambda x: x == 1, res)) if success: ans_acc += 1 if eval_mode or format_valid: reward += acc_score if not eval_mode: reward = reward + length_reward except Exception: pass # Check if the sequence is over length if not eval_mode and res_length >= max_new_tokens: reward *= 0.0 except Exception: pass if not eval_mode: return torch.tensor([reward, format_acc, ans_acc]).to(input_ids.device) else: prompt = tokenizer.decode(input_ids[:s], skip_special_tokens=True) return { "prompt": prompt, "prediction": decoded_final_answer, "test_cases": test_cases, "test_results": res, "test_metadata": metadata, "parsed": solution, "format_valid": format_acc.item(), "ans_valid": ans_acc.item(), "response_length": res_length, "reward": reward.item(), }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/reward/verifiable_reward.py
applications/ColossalChat/coati/distributed/reward/verifiable_reward.py
""" Function-based reward verification module. """ import inspect from typing import Any, Dict, List import torch class VerifiableReward: def __init__(self, reward_fns: List[callable], **kwargs: List[Dict[str, Any]]): self.reward_fns = reward_fns self.kwargs = kwargs def __call__( self, input_ids: torch.LongTensor, gt_answer: List[str] = None, test_cases: List[str] = None, response_idx: List[torch.Tensor] = None, ) -> torch.Tensor: # Get batch size bs = input_ids.size(0) # Initialize reward rewards = torch.zeros((bs, 3), device=input_ids.device) # Loop through reward functions for reward_fn in self.reward_fns: # Apply the reward function to the entire batch at once if "gt_answer" in inspect.getfullargspec(reward_fn).args: reward_batch = torch.stack( [ reward_fn( input_ids[i], gt_answer=gt_answer[i], response_idx=response_idx[i], **self.kwargs, ) for i in range(bs) ], dim=0, ) elif "test_cases" in inspect.getfullargspec(reward_fn).args: reward_batch = torch.stack( [ reward_fn( input_ids[i], test_cases=test_cases[i], response_idx=response_idx[i], **self.kwargs, ) for i in range(bs) ], dim=0, ) else: reward_batch = torch.stack( [ reward_fn( input_ids[i], response_idx=response_idx[i], **self.kwargs, ) for i in range(bs) ], dim=0, ) rewards += reward_batch return rewards
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/reward/code_reward/testing_util.py
applications/ColossalChat/coati/distributed/reward/code_reward/testing_util.py
# Code from the verl Project (https://github.com/agentica-project/rllm), # which itself is adapted from Prime (https://github.com/PRIME-RL/PRIME) # # Copyright 2024 ByteDance Group # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import ast import faulthandler import json import platform # to run the solution files we're using a timing based approach import signal import sys import traceback # used for debugging to time steps from datetime import datetime from enum import Enum # for capturing the stdout from io import StringIO # used for testing the code that reads from input from unittest.mock import mock_open, patch import numpy as np from pyext import RuntimeModule def truncatefn(s, length=300): assert isinstance(s, str) if len(s) <= length: return s return s[: length // 2] + "...(truncated) ..." + s[-length // 2 :] class CODE_TYPE(Enum): call_based = 0 standard_input = 1 # used to capture stdout as a list # from https://stackoverflow.com/a/16571630/6416660 # alternative use redirect_stdout() from contextlib class Capturing(list): def __enter__(self): self._stdout = sys.stdout sys.stdout = self._stringio = StringIO() # Make closing the StringIO a no-op self._stringio.close = lambda x: 1 return self def __exit__(self, *args): self.append(self._stringio.getvalue()) del self._stringio # free up some memory sys.stdout = self._stdout def only_int_check(val): return isinstance(val, int) def string_int_check(val): return isinstance(val, str) and val.isdigit() def combined_int_check(val): return only_int_check(val) or string_int_check(val) def clean_traceback(error_traceback): file_start = error_traceback.find('File "<string>"') # print(file_start) error_traceback = "Traceback (most recent call last):\n " + error_traceback[file_start:] return error_traceback def run_test(in_outs, test=None, debug=False, timeout=15, run_all_tests=False): """ if test(generated_code) is not None it'll try to run the code. otherwise it'll just return an input and output pair. """ # Disable functionalities that can make destructive changes to the test. reliability_guard() if debug: print(f"start = {datetime.now().time()}") if in_outs: if in_outs.get("fn_name") is None: which_type = CODE_TYPE.standard_input # Standard input method_name = None else: which_type = CODE_TYPE.call_based # Call-based method_name = in_outs["fn_name"] if debug: print(f"loaded input_output = {datetime.now().time()}") if test is None: raise AssertionError("should not happen: test code is none") elif test is not None: results = [] sol = "from string import *\nfrom re import *\nfrom datetime import *\nfrom collections import *\nfrom heapq import *\nfrom bisect import *\nfrom copy import *\nfrom math import *\nfrom random import *\nfrom statistics import *\nfrom itertools import *\nfrom functools import *\nfrom operator import *\nfrom io import *\nfrom sys import *\nfrom json import *\nfrom builtins import *\nfrom typing import *\nimport string\nimport re\nimport datetime\nimport collections\nimport heapq\nimport bisect\nimport copy\nimport math\nimport random\nimport statistics\nimport itertools\nimport functools\nimport operator\nimport io\nimport sys\nimport json\nsys.setrecursionlimit(6*10**5)\n" # noqa: E501 if debug: print(f"loading test code = {datetime.now().time()}") if which_type == CODE_TYPE.call_based: sol += test if debug: print(f"sol = {sol}") signal.alarm(timeout) try: tmp_sol = RuntimeModule.from_string("tmp_sol", "", sol) tmp = tmp_sol if "class Solution" not in test else tmp_sol.Solution() signal.alarm(0) except Exception as e: signal.alarm(0) error_traceback = traceback.format_exc() if debug: print(f"type 0 compilation error = {e}") results.append(-2) return results, { "error": repr(e), # "error_code": -1, # "error_message": "Compilation Error", "traceback": clean_traceback(error_traceback), } signal.alarm(0) elif which_type == CODE_TYPE.standard_input: # sol # if code has if __name__ == "__main__": then remove it try: astree = ast.parse(test) last_block = astree.body[-1] if isinstance(last_block, ast.If): condition = last_block.test if ast.unparse(condition).strip() == "__name__ == '__main__'": test = ast.unparse(astree.body[:-1]) + "\n" + ast.unparse(last_block.body) except Exception: pass tmp_test = test.split("\n") new_test = [] for x in tmp_test: if (not x.startswith("from ")) and (not x.startswith("import ")): new_test.append("\t" + x + "\n") else: new_test.append(x + "\n") tmp_test = new_test new_test = "" started = False for i in tmp_test: if i.startswith("\t") and not started: new_test += "stdin = sys.stdin\nstdout = sys.stdout\n" new_test += "def code():\n" new_test += i started = True elif started and ((i.startswith("from ")) or (i.startswith("import "))): new_test += "\t" + i else: new_test += i tmp_test = new_test sol += tmp_test method_name = "code" signal.alarm(timeout) try: tmp_sol = RuntimeModule.from_string("tmp_sol", "", sol) tmp = tmp_sol signal.alarm(0) except Exception as e: signal.alarm(0) error_traceback = traceback.format_exc() if debug: print(f"type 1 compilation error = {e}") results.append(-2) return results, { "error": repr(e), # "error_code": -1, # "error_message": "Compilation Error", "traceback": clean_traceback(error_traceback), } signal.alarm(0) if debug: print(f"get method {method_name} = {datetime.now().time()}") try: method = getattr(tmp, method_name) # get_attr second arg must be str except Exception: signal.alarm(0) error_traceback = traceback.format_exc() error_info = sys.exc_info() print(f"unable to get function error = {error_info}") results.append(-2) return results, { "error": repr(error_info), # "error_code": -1, # "error_message": "Unable to extract code", "traceback": clean_traceback(error_traceback), } for index, inputs in enumerate(in_outs["inputs"]): raw_inputs = inputs raw_outputs = in_outs["outputs"][index] if which_type == CODE_TYPE.call_based: inputs = [json.loads(line) for line in inputs.split("\n")] in_outs["outputs"][index] = json.loads(in_outs["outputs"][index]) truncate_line_size = 300 // (raw_inputs.count("\n") + 1) raw_inputs = "\n".join( [truncatefn(line, truncate_line_size) for line in raw_inputs.strip().split("\n")] ) raw_outputs = truncatefn(raw_outputs, 200) else: raw_inputs = truncatefn(raw_inputs) raw_outputs = truncatefn(raw_outputs, 200) # JSON forces dictionaries to have string keys; this undoes this (assuming a singleton list) try: if isinstance(inputs[0], dict): inputs = [{int(k): v for k, v in inputs[0].items()}] except Exception: pass try: if isinstance(in_outs["outputs"][index], dict): in_outs["outputs"][index] = [{int(k): v for k, v in in_outs["outputs"][index].items()}] except Exception: pass try: if isinstance(in_outs["outputs"][index][0], dict): in_outs["outputs"][index] = [{int(k): v for k, v in in_outs["outputs"][index][0].items()}] except Exception: pass if debug: print( f"time: {datetime.now().time()} testing index = {index} inputs = {inputs}, {type(inputs)}. type = {which_type}" ) if which_type == CODE_TYPE.call_based: # Call-based signal.alarm(timeout) faulthandler.enable() try: output = method(*inputs) raw_true_output = output raw_true_output_copy = json.dumps(output) raw_true_output_copy = truncatefn(raw_true_output_copy, 200) # ground truth sequences are not tuples if isinstance(output, tuple): output = list(output) tmp_result = output == in_outs["outputs"][index] if isinstance(in_outs["outputs"][index], list) and in_outs["outputs"][index]: tmp_result = tmp_result or (output == in_outs["outputs"][index][0]) # ground truth sequences are not tuples try: if isinstance(output[0], tuple): tmp_result = tmp_result or ([list(x) for x in output] == in_outs["outputs"][index][0]) except Exception: pass results.append(tmp_result) if tmp_result is not True: return results, { "output": raw_true_output_copy, "expected": raw_outputs, "inputs": raw_inputs, # "error_code": -2, "error_message": "Wrong Answer", } # reset the alarm signal.alarm(0) except Exception as e: signal.alarm(0) error_traceback = traceback.format_exc() faulthandler.disable() if debug: print(f"Standard input runtime error or time limit exceeded error = {e}") results.append(-1) return results, { "error": repr(e), "traceback": clean_traceback(error_traceback), } faulthandler.disable() signal.alarm(0) if debug: print( f"outputs = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs}, {type(inputs)}, {output == [in_outs['outputs'][index]]}" ) elif which_type == CODE_TYPE.standard_input: # Standard input faulthandler.enable() passed = False if isinstance(inputs, list): inputs = "\n".join(inputs) if isinstance(in_outs["outputs"][index], list): in_outs["outputs"][index] = "\n".join(in_outs["outputs"][index]) signal.alarm(timeout) with Capturing() as output: try: call_method(method, inputs) # reset the alarm signal.alarm(0) passed = True except Exception as e: # runtime error or took too long signal.alarm(0) error_traceback = traceback.format_exc() print(f"Call-based runtime error or time limit exceeded error = {repr(e)}{e}") results.append(-1) signal.alarm(0) if run_all_tests: continue return results, { "error": repr(e), "traceback": clean_traceback(error_traceback), } signal.alarm(0) raw_true_output = output[0] raw_true_output_copy = truncatefn(raw_true_output, 200) output = raw_true_output.splitlines() if not passed: if debug: nl = "\n" if not isinstance(inputs, list): print( f"not passed output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs.replace(nl, ' new-line ')}, {type(inputs)}, {output == [in_outs['outputs'][index]]}" ) else: print( f"not passed output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs}, {type(inputs)}, {output == [in_outs['outputs'][index]]}" ) continue if passed and debug: print(f"==> output = {output}, test outputs = {in_outs['outputs'][index]}") if custom_compare_(output, in_outs["outputs"][index]): tmp_result = True results.append(tmp_result) continue # ground truth sequences are expressed as lists not tuples if isinstance(output, tuple): output = list(output) tmp_result = False try: tmp_result = output == [in_outs["outputs"][index]] if isinstance(in_outs["outputs"][index], list): tmp_result = tmp_result or (output == in_outs["outputs"][index]) if isinstance(output[0], str): tmp_result = tmp_result or ([e.strip() for e in output] == in_outs["outputs"][index]) except Exception as e: if debug: print(f"Failed check1 exception = {e}") if tmp_result is True: results.append(tmp_result) continue # try one more time without \n if isinstance(in_outs["outputs"][index], list): for tmp_index, i in enumerate(in_outs["outputs"][index]): in_outs["outputs"][index][tmp_index] = i.split("\n") in_outs["outputs"][index][tmp_index] = [ x.strip() for x in in_outs["outputs"][index][tmp_index] if x ] else: in_outs["outputs"][index] = in_outs["outputs"][index].split("\n") in_outs["outputs"][index] = list(filter(len, in_outs["outputs"][index])) in_outs["outputs"][index] = list(map(lambda x: x.strip(), in_outs["outputs"][index])) try: tmp_result = output == [in_outs["outputs"][index]] if isinstance(in_outs["outputs"][index], list): tmp_result = tmp_result or (output == in_outs["outputs"][index]) except Exception as e: if debug: print(f"Failed check2 exception = {e}") if tmp_result is True: results.append(tmp_result) continue # try by converting the output into a split up list too if isinstance(output, list): output = list(filter(len, output)) if debug: nl = "\n" if not isinstance(inputs, list): print( f"@1 output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs.replace(nl, ' new-line ')}, {type(inputs)}, {output == [in_outs['outputs'][index]]} {tmp_result=}" ) else: print( f"@1 output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs}, {type(inputs)}, {output == [in_outs['outputs'][index]]} {tmp_result=}" ) if debug: print(f"{tmp_result=} @a") try: tmp_result = output == [in_outs["outputs"][index]] if isinstance(in_outs["outputs"][index], list): tmp_result = tmp_result or (output == in_outs["outputs"][index]) except Exception as e: if debug: print(f"Failed check3 exception = {e}") if debug: print(f"{tmp_result=} @b") try: all_ints = all( combined_int_check(e1) and combined_int_check(e2) for e1, e2 in zip(output, in_outs["outputs"][index]) ) if not all_ints: if debug: print( [ combined_int_check(e1) and combined_int_check(e2) for e1, e2 in zip(output, in_outs["outputs"][index]) ] ) output_float = [float(e) for e in output] gt_float = [float(e) for e in in_outs["outputs"][index]] tmp_result = tmp_result or ( (len(output_float) == len(gt_float)) and np.allclose(output_float, gt_float) ) except Exception: pass if debug: print(f"{tmp_result=} @c") try: if isinstance(output[0], list): all_ints = all( combined_int_check(e1) and combined_int_check(e2) for e1, e2 in zip(output[0], in_outs["outputs"][index]) ) if not all_ints: output_float = [float(e) for e in output[0]] gt_float = [float(e) for e in in_outs["outputs"][index][0]] tmp_result = tmp_result or ( (len(output_float) == len(gt_float)) and np.allclose(output_float, gt_float) ) except Exception: pass if tmp_result is True: results.append(tmp_result) continue if debug: print(f"{tmp_result=} @d") # try by converting the stuff into split up list if isinstance(in_outs["outputs"][index], list): for tmp_index, i in enumerate(in_outs["outputs"][index]): in_outs["outputs"][index][tmp_index] = set(i.split()) else: in_outs["outputs"][index] = set(in_outs["outputs"][index].split()) if debug: print(f"{tmp_result=} @e") try: tmp_result = output == in_outs["outputs"][index] except Exception as e: if debug: print(f"Failed check4 exception = {e}") continue if tmp_result is True: results.append(tmp_result) continue if debug: print(f"{tmp_result=} @f") # try by converting the output into a split up list too if isinstance(output, list): for tmp_index, i in enumerate(output): output[tmp_index] = i.split() output = list(filter(len, output)) for tmp_index, i in enumerate(output): output[tmp_index] = set(i) else: output = output.split() output = list(filter(len, output)) output = set(output) if debug: print(f"{tmp_result=} @g") if tmp_result is True and debug: print("PASSED") results.append(tmp_result) if tmp_result is not True: if debug: print("final result:", results) if run_all_tests: continue return results, { "output": raw_true_output_copy, "expected": raw_outputs, "inputs": raw_inputs, # "error_code": -2, "error_message": "Wrong Answer", } if debug: nl = "\n" if not isinstance(inputs, list): print( f"@2 output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs.replace(nl, ' new-line ')}, {type(inputs)}, {output == [in_outs['outputs'][index]]}" ) else: print( f"@2 output = {output}, test outputs = {in_outs['outputs'][index]}, inputs = {inputs}, {type(inputs)}, {output == [in_outs['outputs'][index]]}" ) print(f"results = {results}") if debug: print("final results", results) return results, {} def custom_compare_(output, ground_truth): if isinstance(output, list): output_1 = "\n".join(output) if stripped_string_compare(output_1, ground_truth): return True if isinstance(output, list): output_2 = [o.lstrip().rstrip() for o in output] output_2 = "\n".join(output_2) if stripped_string_compare(output_2, ground_truth): return True return False def stripped_string_compare(s1, s2): s1 = s1.lstrip().rstrip() s2 = s2.lstrip().rstrip() return s1 == s2 def call_method(method, inputs): if isinstance(inputs, list): inputs = "\n".join(inputs) inputs_line_iterator = iter(inputs.split("\n")) # sys.setrecursionlimit(10000) # @patch('builtins.input', side_effect=inputs.split("\n")) @patch("builtins.open", mock_open(read_data=inputs)) @patch("sys.stdin", StringIO(inputs)) @patch("sys.stdin.readline", lambda *args: next(inputs_line_iterator)) @patch("sys.stdin.readlines", lambda *args: inputs.split("\n")) @patch("sys.stdin.read", lambda *args: inputs) # @patch('sys.stdout.write', print) def _inner_call_method(_method): try: return _method() except SystemExit: pass finally: pass return _inner_call_method(method) def reliability_guard(maximum_memory_bytes=None): """ This disables various destructive functions and prevents the generated code from interfering with the test (e.g. fork bomb, killing other processes, removing filesystem files, etc.) WARNING This function is NOT a security sandbox. Untrusted code, including, model- generated code, should not be blindly executed outside of one. See the Codex paper for more information about OpenAI's code sandbox, and proceed with caution. """ if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS, (maximum_memory_bytes, maximum_memory_bytes)) resource.setrlimit(resource.RLIMIT_DATA, (maximum_memory_bytes, maximum_memory_bytes)) if platform.uname().system != "Darwin": resource.setrlimit(resource.RLIMIT_STACK, (maximum_memory_bytes, maximum_memory_bytes)) faulthandler.disable() import builtins builtins.exit = None builtins.quit = None import os os.environ["OMP_NUM_THREADS"] = "1" os.kill = None os.system = None # 防止干扰repl评测 os.putenv = None os.remove = None os.removedirs = None os.rmdir = None os.fchdir = None os.setuid = None os.fork = None os.forkpty = None os.killpg = None os.rename = None os.renames = None os.truncate = None os.replace = None os.unlink = None os.fchmod = None os.fchown = None os.chmod = None os.chown = None os.chroot = None os.lchflags = None os.lchmod = None os.lchown = None os.getcwd = None os.chdir = None import shutil shutil.rmtree = None shutil.move = None shutil.chown = None import subprocess subprocess.Popen = None # type: ignore __builtins__["help"] = None import sys sys.modules["ipdb"] = None sys.modules["joblib"] = None sys.modules["resource"] = None sys.modules["psutil"] = None sys.modules["tkinter"] = None
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/reward/code_reward/utils.py
applications/ColossalChat/coati/distributed/reward/code_reward/utils.py
# Code from the verl Project (https://github.com/agentica-project/rllm), # which itself is adapted from Prime (https://github.com/PRIME-RL/PRIME) # # Copyright 2024 ByteDance Group # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import multiprocessing import traceback from typing import Optional import requests from .testing_util import run_test def _temp_run(sample, generation, debug, result, metadata_list, timeout): try: res, metadata = run_test(in_outs=sample, test=generation, debug=debug, timeout=timeout) result.append(res) metadata_list.append(metadata) except Exception: # print(e) # some tracebacks are extremely long. traceback.print_exc(10) result.append([-1 for i in range(len(sample["inputs"]))]) metadata_list.append({}) def check_correctness(in_outs: Optional[dict], generation, timeout=10, debug=True): """Check correctness of code generation with a global timeout. The global timeout is to catch some extreme/rare cases not handled by the timeouts inside `run_test`""" manager = multiprocessing.Manager() result = manager.list() metadata_list = manager.list() p = multiprocessing.Process(target=_temp_run, args=(in_outs, generation, debug, result, metadata_list, timeout)) p.start() p.join(timeout=600) # Global timeout of 10 minutes that's for all test cases combined if p.is_alive(): p.kill() # p.terminate() if not result: # consider that all tests failed result = [[-1 for i in range(len(in_outs["inputs"]))]] if debug: print("global timeout") return result[0], metadata_list def check_correctness_code_api( in_outs: Optional[dict], generation, timeout=10, debug=True, url="http://localhost:8000/check_correctness" ): payload = {"in_outs": in_outs, "generation": generation, "timeout": timeout, "debug": debug} response = requests.post(url, json=payload) if response.status_code == 200: results = response.json() return results["result"], results["metadata"] else: print(f"Error: {response.status_code} - {response.text}") return [-1 for i in range(len(in_outs["inputs"]))], {}
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/zero_bubble/consumer.py
applications/ColossalChat/coati/distributed/zero_bubble/consumer.py
import os import threading import time from typing import Any, Dict, Optional import ray import ray.util.collective as cc import torch import torch.distributed as dist from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict from coati.distributed.profiling_utils import CustomProfiler from coati.distributed.utils import bind_batch, post_recv, unbind_batch from tqdm import tqdm from colossalai.booster import Booster from colossalai.booster.plugin import HybridParallelPlugin from colossalai.initialize import launch from colossalai.utils import get_current_device class BaseConsumer: def __init__( self, shared_sync_data_actor: SharedVariableActor, shared_signal_actor: SharedVariableActor, num_producers: int, num_episodes: int, rank: int, world_size: int, master_addr: str, master_port: int, train_dataset_size: int, batch_size: int, model_config: Dict[str, Any], plugin_config: Dict[str, Any], minibatch_size: int = 1, save_interval: int = 100, save_dir: str = "./model", enable_profiling: bool = False, ): self.num_producers = num_producers self.num_episodes = num_episodes self.rank = rank self.world_size = world_size self.master_addr = master_addr self.master_port = master_port self.train_dataset_size = train_dataset_size self.received_prompts = 0 self.batch_size = batch_size self.minibatch_size = minibatch_size self.save_interval = save_interval self.save_dir = save_dir self.enable_profiling = enable_profiling assert batch_size % minibatch_size == 0, "batch_size should be divisible by microbatch_size" self.num_microbatches = batch_size // minibatch_size self.data_uid = 0 self.sync_model_thread_started = False self.model_config = model_config self.plugin_config = plugin_config self.device = get_current_device() self.lr_scheduler = None self.shared_sync_data_actor = shared_sync_data_actor self.shared_signal_actor = shared_signal_actor self.state_dict_cpu = {} def setup(self) -> None: launch(self.rank, self.world_size, self.master_addr, self.master_port, local_rank=0) plugin_config = dict(tp_size=1, pp_size=1, precision="bf16", zero_stage=2) if ( self.plugin_config.get("pp_size", 1) > 1 and "num_microbatches" not in self.plugin_config and "microbatch_size" not in self.plugin_config ): plugin_config["microbatch_size"] = max(1, self.minibatch_size // plugin_config.get("pp_size", 1)) plugin_config.update(self.plugin_config) self.plugin = HybridParallelPlugin(**plugin_config) self.booster = Booster(plugin=self.plugin) self.dp_rank = dist.get_rank(self.plugin.dp_group) self.tp_rank = dist.get_rank(self.plugin.tp_group) self.pp_rank = dist.get_rank(self.plugin.pp_group) self.dp_size = dist.get_world_size(self.plugin.dp_group) self.tp_size = dist.get_world_size(self.plugin.tp_group) self.pp_size = dist.get_world_size(self.plugin.pp_group) self.buffer = [] self.recv_cnt = 0 self.profiler = CustomProfiler(f"C{self.rank}", disabled=not self.enable_profiling) def get_ddp_config(self) -> Dict[str, Any]: """ Get the DDP configuration for the consumer. This method is used to get the DDP configuration for the consumer. """ return { "dp_size": self.dp_size, "tp_size": self.tp_size, "pp_size": self.pp_size, "dp_rank": self.dp_rank, "tp_rank": self.tp_rank, "pp_rank": self.pp_rank, "world_size": self.world_size, "rank": self.rank, } def init_collective_group( self, world_size: int, rank: int, backend: str = "nccl", group_name: str = "default", gloo_timeout: int = 3000000, ): cc.init_collective_group( world_size=world_size, rank=rank, backend=backend, group_name=group_name, gloo_timeout=gloo_timeout ) print(f"[C{self.rank}] Initialized {group_name} collective group", flush=True) def state_dict(self) -> Dict[str, torch.Tensor]: raise NotImplementedError def step(self, **kwargs) -> Optional[float]: raise NotImplementedError def prepare_mini_batch(self, effective_group_to_raw_group_mapping: Dict[int, int]) -> Dict[str, torch.Tensor]: """ Prepare a mini-batch from the effective group to raw group mapping. This method is used to create a mini-batch for training. """ batches = [ self.buffer[effective_group_to_raw_group_mapping[i]] for i in range(self.dp_rank * self.minibatch_size, (self.dp_rank + 1) * self.minibatch_size) ] # every dp_rank will receive a complete mini-batch, no need to sync within step() later # each mini-batch use the first self.dp_size * minibatch_size effective samples raw_mini_batches = self.buffer[ : effective_group_to_raw_group_mapping[self.dp_size * self.minibatch_size - 1] + 1 ] # include the last effective sample raw_mini_batches_metric_dict = { "raw_train_mini_batch_reward": [t[1] for t in raw_mini_batches], "raw_train_mini_batch_format_acc": [t[2] for t in raw_mini_batches], "raw_train_mini_batch_ans_acc": [t[3] for t in raw_mini_batches], "raw_train_mini_batch_response_len": [t[4] for t in raw_mini_batches], } batch = bind_batch([t[0] for t in batches]) batch = post_recv(batch) return batch, raw_mini_batches_metric_dict def calculate_effective_group_to_raw_group_mapping(self): effective_group_to_raw_group_mapping = {} for buffer_idx in range(len(self.buffer)): if self.buffer[buffer_idx][0] is not None: effective_group_to_raw_group_mapping[len(effective_group_to_raw_group_mapping)] = buffer_idx return effective_group_to_raw_group_mapping def loop(self) -> None: print(f"Consumer{self.rank}, nmb: {self.num_microbatches}") for episode in range(self.num_episodes): with tqdm( range(self.train_dataset_size), desc=f"Episode {episode} with rollout step(s)", disable=self.rank != 0, ) as pbar: while self.received_prompts < self.train_dataset_size: torch.cuda.reset_peak_memory_stats() effective_group_to_raw_group_mapping = {} self.profiler.enter(f"recv_data") while len(effective_group_to_raw_group_mapping) < self.dp_size * self.minibatch_size: # receive data from producers raw_batch = ray.get( self.shared_sync_data_actor.get_data.remote(self.data_uid) ) # get the first queued data self.profiler.log(f"enter sleep") while raw_batch is None: print( f"[T{dist.get_rank()}] No data received by consumer {self.rank}, skipping. Consider increasing the data actor buffer limit" ) time.sleep(1) raw_batch = ray.get(self.shared_sync_data_actor.get_data.remote(self.data_uid)) continue self.profiler.log(f"exit sleep") self.data_uid += 1 raw_batch = {k: v.to(self.device) for k, v in raw_batch.items()} # calculate group reward et al. filtering. As only the filtered group will be used for training (which is incomplete), # we need to calculate the metrics before filtering here for logging # [batch_size, num_generations] -> [batch_size] reward = raw_batch["reward"][:, :, 0] format_acc = raw_batch["format_acc"][:, :, 0] ans_acc = raw_batch["ans_acc"][:, :, 0] response_len = ( raw_batch["response_idx"][:, :, 1] - raw_batch["response_idx"][:, :, 0] + 1 ).type(torch.float32) effective_group_mask = None if self.filter_range is not None and self.grpo_config.get("dynamic_batching", True): # filter the group based on the reward and accuracy group_ans_acc_mean = ans_acc.mean(dim=1) effective_group_mask = torch.logical_and( group_ans_acc_mean > self.filter_range[0], group_ans_acc_mean < self.filter_range[1] ) raw_batch = unbind_batch(raw_batch) # List[Dict[str, torch.Tensor]] self.received_prompts += len(raw_batch) pbar.update(len(raw_batch)) for group_idx, group_with_reward in enumerate(raw_batch): self.buffer.append( [ ( group_with_reward if effective_group_mask is None or effective_group_mask[group_idx] else None ), reward[group_idx], format_acc[group_idx], ans_acc[group_idx], response_len[group_idx], ] ) if effective_group_mask is not None: print( f"[T{dist.get_rank()}] Filter recv data: {len(raw_batch)} -> {torch.sum(effective_group_mask).cpu().item()} effective groups" ) # mapping the effective group to the raw group for indexing effective_group_to_raw_group_mapping = self.calculate_effective_group_to_raw_group_mapping() print( f"[T{dist.get_rank()}] Collect Effective Prompt: {len(effective_group_to_raw_group_mapping)}/{self.dp_size * self.minibatch_size}" ) self.profiler.exit(f"recv_data") need_sync_model = False while len(effective_group_to_raw_group_mapping) >= self.dp_size * self.minibatch_size: # after we have enough effective groups, we can start training # on each dp_rank, we use minibatch_size effective samples to form a batch batch, raw_mini_batches_metric_dict = self.prepare_mini_batch( effective_group_to_raw_group_mapping ) self.profiler.enter("step") loss = self.step(pbar, **batch, **raw_mini_batches_metric_dict) self.profiler.exit("step") self.buffer = self.buffer[ effective_group_to_raw_group_mapping[self.dp_size * self.minibatch_size - 1] + 1 : ] # recalculate the effective group to raw group mapping effective_group_to_raw_group_mapping_size_before = len(effective_group_to_raw_group_mapping) effective_group_to_raw_group_mapping = self.calculate_effective_group_to_raw_group_mapping() assert ( len(effective_group_to_raw_group_mapping) == effective_group_to_raw_group_mapping_size_before - self.dp_size * self.minibatch_size ) # cc.barrier(group_name="consumer_pg") if loss is not None: pbar.set_postfix({"loss": loss}) need_sync_model = True ray.get(self.shared_signal_actor.set_signal.remote("global_step", self.global_step + 1)) if need_sync_model and ( (self.global_step + 1) % self.save_interval == 0 or self.received_prompts >= self.train_dataset_size ): if self.rank == 0: print(f"Start saving policy model at step {self.global_step + 1}.") save_path = os.path.join( self.save_dir, f"modeling-episode-{episode}-step-{self.global_step + 1}" ) self.booster.save_model(self.policy_model, save_path, shard=True) if self.rank == 0: print(f"Saved model checkpoint at step {self.global_step + 1} in folder {save_path}") if need_sync_model and ( episode != self.num_episodes - 1 or self.received_prompts != self.train_dataset_size ): def sync_model_thread(): # sync model weights to all producers, if no model update or it is the last training step, skip syncing if self.pp_size > 1: print( f"[T{dist.get_rank()}] Sync model PP stage {self.pp_rank} episode {episode} step {self.global_step}" ) else: print(f"[T{dist.get_rank()}] Sync model episode {episode} step {self.global_step}") torch.cuda.empty_cache() if self.pp_size > 1: if self.tp_rank == 0 and self.dp_rank == 0: self.profiler.enter("sync_model") ray.get( self.shared_signal_actor.set_signal.remote( f"consumer_pp_{self.pp_rank}", "ready_sync_model" ) ) print( f"[T{dist.get_rank()}] Sync model PP stage {self.pp_rank} episode {episode} step {self.global_step}" ) ray_broadcast_tensor_dict( self.state_dict_cpu, src=0, device=torch.device("cpu"), group_name=f"sync_model_consumer_pp_{self.pp_rank}", backend="gloo", ) self.profiler.exit("sync_model") else: if self.rank == 0: self.profiler.enter("sync_model") ray.get(self.shared_signal_actor.set_signal.remote("consumer", "ready_sync_model")) print(f"[T{dist.get_rank()}] Sync model episode {episode} step {self.global_step}") ray_broadcast_tensor_dict( self.state_dict_cpu, src=0, device=torch.device("cpu"), group_name="sync_model_consumer", backend="gloo", ) self.profiler.exit("sync_model") if not self.sync_model_thread_started: # only sync model when the thread is not started and no other thread is broadcasting self.sync_model_thread_started = True state_dict_ = self.state_dict() if (self.pp_size > 1 and self.tp_rank == 0 and self.dp_rank == 0) or ( self.pp_size == 1 and self.rank == 0 ): if len(self.state_dict_cpu) == 0: # use pinned memory to speed up the transfer self.state_dict_cpu = {k: v.cpu().pin_memory() for k, v in state_dict_.items()} torch.cuda.synchronize() for k, v in state_dict_.items(): self.state_dict_cpu[k].copy_(v, non_blocking=True) torch.cuda.synchronize() cc.barrier( group_name="consumer_pg" ) # to make sure all ranks have state dict offloaded to CPU before starting the thread time_before_starting_thread = time.time() threading.Thread(target=sync_model_thread).start() # sync_model_thread() self.profiler.log( f"Sync model, took {time.time() - time_before_starting_thread:.2f} seconds" ) self.sync_model_thread_started = False # ray.get(self.shared_signal_actor.release_process_lock.remote("broadcasting_lock")) self.profiler.log(f"Peak memory usage: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") self.received_prompts = 0 ray.get(self.shared_signal_actor.set_signal.remote("consumer", "terminate")) def __del__(self): if hasattr(self, "profiler"): self.profiler.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/zero_bubble/producer.py
applications/ColossalChat/coati/distributed/zero_bubble/producer.py
import copy import json import os import threading import time from typing import Any, Dict, Optional import ray import ray.util.collective as cc import torch import tqdm import wandb from coati.dataset.loader import RawConversationDataset, collate_fn_grpo from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict from coati.distributed.inference_backend import BACKEND_MAP from coati.distributed.profiling_utils import CustomProfiler from coati.distributed.reward.reward_fn import boxed_math_reward_fn, code_reward_fn, math_reward_fn from coati.distributed.reward.verifiable_reward import VerifiableReward from coati.distributed.utils import pre_send, safe_append_to_jsonl_file from ray.util.collective import allreduce from ray.util.collective.types import ReduceOp from torch.utils.data import DataLoader, DistributedSampler from transformers import AutoTokenizer from colossalai.utils import get_current_device try: from vllm import SamplingParams except ImportError: LLM = None class BaseProducer: def __init__( self, shared_sync_data_actor: SharedVariableActor, shared_signal_actor: SharedVariableActor, producer_idx: int, num_producers: int, num_consumer_procs: int, num_episodes: int, batch_size: int, train_dataset_config: Dict[str, Any], model_config: Dict[str, Any], generate_config: Dict[str, Any], tokenizer_config: Optional[Dict[str, Any]] = None, microbatch_size: int = 1, backend: str = "transformers", consumer_plugin_config: Dict[str, Any] = None, eval_dataset_config=None, eval_interval=-1, # disable evaluation grpo_config: Dict[str, Any] = None, eval_save_dir: str = "./eval", project_name: str = None, run_name: str = None, wandb_group_name: str = None, log_rollout_interval: int = 20, rollout_log_file: str = "./rollout_log.jsonl", enable_profiling: bool = False, ): self.producer_idx = producer_idx self.num_producers = num_producers self.num_consumer_procs = num_consumer_procs self.num_episodes = num_episodes self.batch_size = batch_size self.microbatch_size = microbatch_size assert batch_size % microbatch_size == 0 self.num_microbatches = batch_size // microbatch_size self.latest_eval_step = -1 self.profiler = CustomProfiler(f"P{self.producer_idx}", disabled=not enable_profiling) # for async data and model sync self.shared_sync_data_actor = shared_sync_data_actor self.shared_signal_actor = shared_signal_actor self.sync_model_thread_started = False self.train_dataset_config = train_dataset_config self.model_config = model_config self.generate_config = generate_config self.tokenizer_config = tokenizer_config self.consumer_plugin_config = consumer_plugin_config self.eval_interval = eval_interval self.eval_save_dir = eval_save_dir self.consumer_global_step = 0 self.producer_weight_version = 0 self.eval_mode = False self.log_rollout_interval = log_rollout_interval self.latest_rollout_log_step = -1 self.grpo_config = grpo_config reward_model_kwargs = { k: v for k, v in grpo_config.items() if k in ["soft_over_length_punishment", "max_new_tokens", "cache_length"] } self.response_format_tags = grpo_config.get("response_format_tags", None) if producer_idx == 0: if os.path.exists(rollout_log_file): raise ValueError( f"Rollout log file {rollout_log_file} already exists. Please delete it or change the name." ) else: os.makedirs(os.path.dirname(rollout_log_file), exist_ok=True) self.rollout_log_file = open(rollout_log_file, "w", encoding="utf8") if self.producer_idx == 0: self.wandb_run = wandb.init( project=project_name, sync_tensorboard=False, dir="./wandb", name=run_name + "_eval", group=wandb_group_name, ) if os.path.exists(self.eval_save_dir) and self.eval_interval > 0: raise ValueError(f"Eval save dir {self.eval_save_dir} already exists. Please delete it or change the name.") # init tokenizer if tokenizer_config is None: tokenizer_path = model_config["path"] self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path) else: tokenizer_path = tokenizer_config.pop("path") self.tokenizer = AutoTokenizer.from_pretrained(tokenizer_path, **tokenizer_config) self.tokenizer.padding_side = "left" # init dataloader train_dataset_path = train_dataset_config.pop("path") self.train_dataset = RawConversationDataset(self.tokenizer, train_dataset_path, **train_dataset_config) self.train_dataloader = DataLoader( self.train_dataset, batch_size=microbatch_size, sampler=DistributedSampler( self.train_dataset, num_replicas=num_producers, rank=producer_idx, shuffle=True, drop_last=True, seed=42, ), num_workers=4, drop_last=True, collate_fn=collate_fn_grpo, ) if grpo_config["reward_fn_type"] == "think_answer_tags": self.evaluation_function = math_reward_fn elif grpo_config["reward_fn_type"] == "boxed": self.evaluation_function = boxed_math_reward_fn elif grpo_config["reward_fn_type"] == "code": self.evaluation_function = code_reward_fn else: raise ValueError(f"Unknown evaluation function type {grpo_config['reward_fn_type']}") self.eval_dataset_config = eval_dataset_config if self.eval_dataset_config is not None: self.eval_dataloaders = {} for eval_task_name in self.eval_dataset_config: eval_dataset_path = eval_dataset_config[eval_task_name].pop("path") eval_dataset = RawConversationDataset( self.tokenizer, eval_dataset_path, **eval_dataset_config[eval_task_name] ) print(f"[P{self.producer_idx}] eval dataset {eval_task_name} size: {len(eval_dataset)}") self.eval_dataloaders[eval_task_name] = DataLoader( eval_dataset, batch_size=microbatch_size, sampler=DistributedSampler( eval_dataset, num_replicas=num_producers, rank=producer_idx, shuffle=False, drop_last=False, seed=42, ), collate_fn=collate_fn_grpo, ) else: print("No eval dataset provided, skip eval") self.device = get_current_device() self.reward_model = VerifiableReward( reward_fns=[self.evaluation_function], # multiple reward functions can be added here tokenizer=self.tokenizer, tags=self.response_format_tags, **reward_model_kwargs, ) # init backend if backend in BACKEND_MAP: self.backend_cls = BACKEND_MAP[backend] else: raise ValueError(f"Unexpected backend {backend}") self.consumer_pp_size = consumer_plugin_config.get("pp_size", 1) # consumer pp size self.state_dict_cpu = {i: None for i in range(self.consumer_pp_size)} def init_collective_group( self, world_size: int, rank: int, backend: str = "nccl", group_name: str = "default", gloo_timeout: int = 3000000, ): cc.init_collective_group( world_size=world_size, rank=rank, backend=backend, group_name=group_name, gloo_timeout=gloo_timeout ) print(f"[P{self.producer_idx}] Initialized {group_name} collective group", flush=True) def rollout(self, input_ids: torch.Tensor, attention_mask: torch.Tensor, **kwargs) -> Dict[str, torch.Tensor]: raise NotImplementedError def load_state_dict(self, state_dict: Dict[str, torch.Tensor]) -> None: raise NotImplementedError def loop(self) -> None: num_update_per_episode = len(self.train_dataloader) // self.num_microbatches num_valid_microbatches = num_update_per_episode * self.num_microbatches print( f"[P{self.producer_idx}] num_valid_microbatches {num_valid_microbatches}, nmb: {self.num_microbatches}, dl: {len(self.train_dataloader)}" ) for episode in range(self.num_episodes): self.train_dataloader.sampler.set_epoch(episode) for i, batch in enumerate(self.train_dataloader): self.profiler.log(f"train episode {episode} batch {i}") if i >= num_valid_microbatches: break self.consumer_global_step = ray.get(self.shared_signal_actor.get_signal.remote()).get("global_step", 0) # sync model first, as the model syncing runs in a separate thread, will not block the main thread # sync model during inference, which takes less than 10s, so that the model can be updated immediately after inference if episode != self.num_episodes - 1 or i != num_valid_microbatches - 1: # don't sync model for last iteration if isinstance(self.model, BACKEND_MAP["vllm"]) and self.model.model_config.get( "enable_sleep_mode", False ): self.model.llm.sleep() # revict KV_cache to avoid OOM torch.cuda.empty_cache() # sync model thread function def sync_model_thread(): if self.consumer_pp_size > 1: self.profiler.enter("sync_model") for pp_idx in range(self.consumer_pp_size): ray.get( self.shared_signal_actor.set_signal.remote( f"producer_{self.producer_idx}_pp_{pp_idx}", "ready_sync_model" ) ) for pp_idx in range(self.consumer_pp_size): print( f"[P{self.producer_idx}] Sync model PP stage {pp_idx} episode {episode} step {(i + 1) // self.num_microbatches - 1}" ) self.state_dict_cpu[pp_idx] = ray_broadcast_tensor_dict( self.state_dict_cpu[pp_idx], 1, device=torch.device("cpu"), group_name=f"sync_model_producer_{self.producer_idx}_pp_{pp_idx}", backend="gloo", # use gloo for CPU communication pin_memory=True, ) self.profiler.exit("sync_model") else: self.profiler.enter("sync_model") ray.get( self.shared_signal_actor.set_signal.remote( f"producer_{self.producer_idx}", "ready_sync_model" ) ) print( f"[P{self.producer_idx}] Sync model episode {episode} step {(i + 1) // self.num_microbatches - 1}" ) time0 = time.time() self.state_dict_cpu[0] = ray_broadcast_tensor_dict( self.state_dict_cpu[0], 1, device=torch.device("cpu"), group_name=f"sync_model_producer_{self.producer_idx}", backend="gloo", # use gloo for CPU communication pin_memory=True, ) self.profiler.log(f"Broadcast model state dict took {time.time() - time0:.2f} seconds") self.profiler.exit("sync_model") self.sync_model_thread_started = False distributor_weight_version = ray.get(self.shared_signal_actor.get_signal.remote()).get( f"distributor_weight_version", 0 ) if ( not self.sync_model_thread_started and distributor_weight_version != self.producer_weight_version ): # only sync model when the thread is not started and global step is changed self.sync_model_thread_started = True self.sync_model_thread = threading.Thread(target=sync_model_thread) self.producer_weight_version = distributor_weight_version self.sync_model_thread.start() torch.cuda.empty_cache() if isinstance(self.model, BACKEND_MAP["vllm"]) and self.model.model_config.get( "enable_sleep_mode", False ): self.model.llm.wake_up() if self.eval_interval > 0 and self.eval_dataset_config is not None: if ( self.consumer_global_step - self.latest_eval_step >= self.eval_interval and self.consumer_global_step > self.latest_eval_step ) or self.latest_eval_step == -1: to_log_msg = {} self.eval_mode = True for eval_task_name in self.eval_dataloaders: if self.producer_idx == 0: print( f"[P{self.producer_idx}] Evaluate model at training step {self.consumer_global_step} on task {eval_task_name}" ) eval_results = [] eval_statistics_tensor = torch.zeros((2,), dtype=torch.float32).to(self.device) for eval_batch in tqdm.tqdm( self.eval_dataloaders[eval_task_name], disable=self.producer_idx != 0 ): eval_outputs = self.rollout(**eval_batch, sample_params=self.eval_sample_params) eval_results = eval_results + [ self.evaluation_function( eval_outputs["input_ids"][m][n], eval_outputs[ ( "test_cases" if self.grpo_config["reward_fn_type"] == "code" else "gt_answer" ) ][m], eval_outputs["response_idx"][m][n], tokenizer=self.tokenizer, eval_mode=True, tags=self.response_format_tags, ) for m in range(eval_outputs["input_ids"].size(0)) for n in range(eval_outputs["input_ids"].size(1)) ] eval_statistics_tensor[0] += len([res for res in eval_results if res["ans_valid"] == 1]) eval_statistics_tensor[1] += len(eval_results) allreduce(eval_statistics_tensor, op=ReduceOp.SUM, group_name="producer_pg") to_log_msg[f"eval/{eval_task_name}"] = ( eval_statistics_tensor[0].item() / eval_statistics_tensor[1].item() ) if self.producer_idx == 0: print( f"[P{self.producer_idx}]: Accuracy on {eval_task_name}: {to_log_msg[f'eval/{eval_task_name}']}" ) # save eval results safe_append_to_jsonl_file( os.path.join( self.eval_save_dir, f"{eval_task_name}_training_step_{self.consumer_global_step}.jsonl", ), eval_results, ) if self.producer_idx == 0: self.wandb_run.log(to_log_msg, step=self.consumer_global_step) self.eval_mode = False self.latest_eval_step = self.consumer_global_step self.profiler.enter("sleep") while not (ray.get(self.shared_sync_data_actor.pickup_rollout_task.remote(self.microbatch_size))): time.sleep(1) self.profiler.exit("sleep") self.profiler.enter("rollout") self.profiler.log(f"rollout batch {i} episode {episode}") # time.sleep(30) # simulate long inference time outputs = self.rollout(**batch) self.profiler.exit("rollout") outputs["temperature"] = torch.tensor( [self.model.generate_config["temperature"]] * outputs["input_ids"].size(0) ).to(outputs["input_ids"].device) bs, num_gen = outputs["input_ids"].size(0), outputs["input_ids"].size(1) self.profiler.enter("calculate_reward") if self.grpo_config["reward_fn_type"] == "code": test_cases = [] for prompt_id in range(bs): test_cases.extend([outputs["test_cases"][prompt_id]] * num_gen) reward_model_output = self.reward_model( outputs["input_ids"].view((-1, outputs["input_ids"].size(-1))), test_cases=test_cases, response_idx=outputs["response_idx"].view((-1, 2)), ) else: gt_answer = [] for prompt_id in range(bs): gt_answer.extend([outputs["gt_answer"][prompt_id]] * num_gen) reward_model_output = self.reward_model( outputs["input_ids"].view((-1, outputs["input_ids"].size(-1))), gt_answer=gt_answer, response_idx=outputs["response_idx"].view((-1, 2)), ) outputs["reward"] = ( torch.tensor([value[0] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) outputs["format_acc"] = ( torch.tensor([value[1] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) outputs["ans_acc"] = ( torch.tensor([value[2] for value in reward_model_output]) .to(outputs["input_ids"].device) .view((bs, num_gen, 1)) ) if "gt_answer" in outputs: outputs.pop("gt_answer") if "test_cases" in outputs: outputs.pop("test_cases") self.profiler.exit("calculate_reward") print(f"[P{self.producer_idx}] Send data {[(k, v.shape) for k, v in outputs.items()]}") outputs = pre_send(outputs) outputs = {k: v.cpu() for k, v in outputs.items()} self.profiler.enter("send_data") ray.get(self.shared_sync_data_actor.append_data.remote(outputs)) self.profiler.exit("send_data") if (i + 1) % self.num_microbatches == 0 and ( episode != self.num_episodes - 1 or i != num_valid_microbatches - 1 ): if not self.sync_model_thread_started: # load state dict, note this should be done in the main thread to avoid race condition for pp_idx in range(self.consumer_pp_size): if self.state_dict_cpu[pp_idx] is not None and self.state_dict_cpu[pp_idx] != {}: self.load_state_dict(self.state_dict_cpu[pp_idx]) # linear annealing for 1 episode, temperature from initial to 0.9 if episode <= 0: ratio = 1 - (len(self.train_dataloader) - i) / len(self.train_dataloader) self.model.generate_config["temperature"] = (1 - ratio) * self.generate_config[ "temperature" ] + ratio * 0.9 if isinstance(self.model, BACKEND_MAP["vllm"]): self.model.sample_params.temperature = (1 - ratio) * self.generate_config[ "temperature" ] + ratio * 0.9 def __del__(self): self.profiler.close() @ray.remote class SimpleProducer(BaseProducer): def __init__( self, shared_sync_data_actor: SharedVariableActor, shared_signal_actor: SharedVariableActor, producer_idx, num_producers, num_consumer_procs, num_episodes, batch_size, train_dataset_config, model_config, generate_config, tokenizer_config=None, microbatch_size=1, backend="transformers", num_generations: int = 8, consumer_plugin_config=None, eval_dataset_config=None, eval_interval=-1, # disable evaluation grpo_config: Dict[str, Any] = None, eval_save_dir: str = "./eval", eval_generation_config={}, project_name: str = None, run_name: str = None, wandb_group_name: str = None, log_rollout_interval: int = 20, rollout_log_file: str = "./rollout_log.jsonl", enable_profiling: bool = False, ): super().__init__( shared_sync_data_actor, shared_signal_actor, producer_idx, num_producers, num_consumer_procs, num_episodes, batch_size, train_dataset_config, model_config, generate_config, copy.deepcopy(tokenizer_config), microbatch_size, backend, consumer_plugin_config, eval_dataset_config=eval_dataset_config, eval_interval=eval_interval, grpo_config=grpo_config, eval_save_dir=eval_save_dir, project_name=project_name, run_name=run_name, wandb_group_name=wandb_group_name, log_rollout_interval=log_rollout_interval, rollout_log_file=rollout_log_file, enable_profiling=enable_profiling, ) print("tokenizer_config", tokenizer_config) self.model = self.backend_cls(model_config, generate_config, self.tokenizer, num_generations, tokenizer_config) self.eval_generation_config = copy.deepcopy(self.model.generate_config) self.eval_generation_config["n"] = 1 # use 1 generation for evaluation self.eval_generation_config.update(eval_generation_config) self.eval_sample_params = SamplingParams(**self.eval_generation_config) @torch.no_grad() def rollout(self, input_ids, attention_mask, **kwargs): rollouts = self.model.generate(input_ids, attention_mask, **kwargs) if self.producer_idx == 0 and not self.eval_mode: if ( self.consumer_global_step - self.latest_rollout_log_step >= self.log_rollout_interval or self.latest_rollout_log_step == -1 ): new_record = ( json.dumps( { "train_step": self.consumer_global_step, "rollout": self.tokenizer.batch_decode( rollouts["input_ids"][:, 0], skip_special_tokens=True ), } ) + "\n" ) self.rollout_log_file.write(new_record) self.rollout_log_file.flush() self.latest_rollout_log_step = self.consumer_global_step return rollouts def __del__(self): if self.producer_idx == 0: self.wandb_run.finish() if hasattr(self, "rollout_log_file"): self.rollout_log_file.close() def load_state_dict(self, state_dict): self.model.load_state_dict(state_dict)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/zero_bubble/distributor.py
applications/ColossalChat/coati/distributed/zero_bubble/distributor.py
import time import ray import ray.util.collective as cc import torch from coati.distributed.comm import SharedVariableActor, ray_broadcast_tensor_dict from coati.distributed.profiling_utils import CustomProfiler from colossalai.utils import get_current_device @ray.remote class Distributor: def __init__( self, distributor_id, consumer_pp_size, num_producers, shared_signal_actor: SharedVariableActor, enable_profiling: bool = True, ): self.distributor_id = distributor_id self.weight_version = [0] * consumer_pp_size self.consumer_pp_size = consumer_pp_size self.state_dict_cpu = {} self.num_producers = num_producers self.shared_signal_actor = shared_signal_actor self.device = get_current_device() self.profiler = CustomProfiler(f"D{self.distributor_id}", disabled=not enable_profiling) def init_collective_group( self, world_size: int, rank: int, backend: str = "nccl", group_name: str = "default", gloo_timeout: int = 3000000, ): cc.init_collective_group( world_size=world_size, rank=rank, backend=backend, group_name=group_name, gloo_timeout=gloo_timeout ) print(f"[D] Initialized {group_name} collective group", flush=True) def loop(self): last_weight_version = self.get_weight_version() while True: time.sleep(1) signal = ray.get(self.shared_signal_actor.get_signal.remote()) if self.consumer_pp_size > 1: if all( [signal.get(f"consumer_pp_{i}", None) == "ready_sync_model" for i in range(self.consumer_pp_size)] ): cc.barrier(group_name="distributor_pg") for i in range(self.consumer_pp_size): self.profiler.enter(f"sync_model_consumer_pp_{i}") ray.get(self.shared_signal_actor.set_signal.remote(f"consumer_pp_{i}", "not_ready_sync_model")) # Broadcast the model state dict from consumer to shared variable actor self.state_dict_cpu[i] = ray_broadcast_tensor_dict( None, 0, device=torch.device("cpu"), group_name=f"sync_model_consumer_pp_{i}", backend="gloo", ) self.profiler.exit(f"sync_model_consumer_pp_{i}") self.weight_version[i] += 1 if all( [ signal.get(f"producer_{self.distributor_id}_pp_{i}", None) == "ready_sync_model" for i in range(self.consumer_pp_size) ] ): for i in range(self.consumer_pp_size): self.profiler.enter(f"sync_model_producer_{self.distributor_id}_pp_{i}") # Broadcast the model state dict to all producers ray.get( self.shared_signal_actor.set_signal.remote( f"producer_{self.distributor_id}_pp_{i}", "not_ready_sync_model" ) ) ray_broadcast_tensor_dict( self.state_dict_cpu[i], 1, device=torch.device("cpu"), group_name=f"sync_model_producer_{self.distributor_id}_pp_{i}", backend="gloo", ) self.profiler.exit(f"sync_model_producer_{self.distributor_id}_pp_{i}") else: if signal.get("consumer", None) == "ready_sync_model": self.profiler.enter("sync_model_consumer") cc.barrier(group_name="distributor_pg") ray.get(self.shared_signal_actor.set_signal.remote("consumer", "not_ready_sync_model")) # Broadcast the model state dict from consumer to shared variable actor self.state_dict_cpu = ray_broadcast_tensor_dict( None, 0, device=torch.device("cpu"), group_name="sync_model_consumer", backend="gloo" ) self.profiler.exit("sync_model_consumer") self.weight_version[0] += 1 if signal.get(f"producer_{self.distributor_id}", None) == "ready_sync_model": self.profiler.enter(f"sync_model_producer_{self.distributor_id}") # Broadcast the model state dict to all producers ray.get( self.shared_signal_actor.set_signal.remote( f"producer_{self.distributor_id}", "not_ready_sync_model" ) ) ray_broadcast_tensor_dict( self.state_dict_cpu, 1, device=torch.device("cpu"), group_name=f"sync_model_producer_{self.distributor_id}", backend="gloo", ) self.profiler.exit(f"sync_model_producer_{self.distributor_id}") if signal.get("consumer", None) == "terminate": self.profiler.log("terminate sync model worker") break if last_weight_version != self.get_weight_version(): last_weight_version = self.get_weight_version() ray.get(self.shared_signal_actor.set_signal.remote("distributor_weight_version", last_weight_version)) def get_weight_version(self): return self.weight_version[0]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/zero_bubble/grpo_consumer.py
applications/ColossalChat/coati/distributed/zero_bubble/grpo_consumer.py
from contextlib import nullcontext from typing import Any, Optional import ray import torch import wandb from coati.distributed.comm import SharedVariableActor from coati.distributed.loss import PolicyLoss from coati.distributed.utils import entropy_from_logits, memory_efficient_logprob from coati.distributed.zero_bubble.consumer import BaseConsumer from coati.trainer.utils import all_reduce_mean, all_reduce_sum from transformers import AutoModelForCausalLM, AutoTokenizer from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam @ray.remote class GRPOConsumer(BaseConsumer): def __init__( self, shared_sync_data_actor: SharedVariableActor, shared_signal_actor: SharedVariableActor, num_producers, num_episodes, rank, world_size, master_addr, master_port, train_dataset_size, batch_size, model_config, plugin_config, minibatch_size=1, num_generations=8, tokenizer_config=None, generate_config=None, grpo_config={}, save_interval: int = 100, save_dir="./model", project_name: str = None, run_name: str = None, wandb_group_name: str = None, enable_profiling: bool = False, ): print(f"Using GRPO config: {grpo_config}") if ( plugin_config.get("pp_size", 1) > 1 and "num_microbatches" not in plugin_config and "microbatch_size" not in plugin_config ): plugin_config["microbatch_size"] = max( 1, grpo_config.get("train_microbatch_size") // plugin_config.get("pp_size", 1) ) super().__init__( shared_sync_data_actor, shared_signal_actor, num_producers, num_episodes, rank, world_size, master_addr, master_port, train_dataset_size, batch_size, model_config, plugin_config, minibatch_size, save_interval=save_interval, save_dir=save_dir, enable_profiling=enable_profiling, ) path = model_config.pop("path") self.policy_model = AutoModelForCausalLM.from_pretrained(path, **model_config) self.policy_model.train() self.policy_model.gradient_checkpointing_enable() self.vocab_size = self.policy_model.config.vocab_size self.optimizer = HybridAdam(self.policy_model.parameters(), lr=grpo_config.get("lr", 1e-6)) self.accum_loss = torch.zeros(1, device=self.device) self.accum_kl = torch.zeros(1, device=self.device) self.accum_entropy = torch.zeros(1, device=self.device) self.accum_advantages = torch.zeros(1, device=self.device) self.raw_train_batch_reward = [] self.raw_train_batch_format_acc = [] self.raw_train_batch_ans_acc = [] self.raw_train_batch_response_len = [] self.accum_count = 0 self.generate_config = generate_config self.grpo_config = grpo_config self.project_name = project_name self.effective_sample_count = 0 self.effective_prompt_count = 0 self.project_name = project_name self.run_name = run_name self.wandb_group_name = wandb_group_name self.policy_loss_fn = PolicyLoss( clip_eps_low=grpo_config.get("clip_eps_low", 0.2), clip_eps_high=grpo_config.get("clip_eps_high", 0.2), beta=grpo_config.get("beta", 0.01), loss_variation=grpo_config.get("loss_variation", "sample_level"), ) # Reference model is initialized from policy model. if self.policy_loss_fn.beta > 0: self.reference_model = AutoModelForCausalLM.from_pretrained(path, **model_config) self.reference_model.eval() if tokenizer_config is not None: path = tokenizer_config.pop("path", None) self.tokenizer = AutoTokenizer.from_pretrained(path, **tokenizer_config) else: self.tokenizer = AutoTokenizer.from_pretrained(path) self.pad_token_id = self.tokenizer.pad_token_id self.num_generations = num_generations self.filter_range = grpo_config.get("filter_range", None) if self.filter_range is not None: assert len(self.filter_range) == 2, "Filter range should have 2 values." self.filter_truncated_response = grpo_config.get("filter_truncated_response", False) if self.filter_truncated_response: self.max_length = 0 if "max_tokens" in self.generate_config: self.max_length = self.generate_config["max_tokens"] elif "max_new_tokens" in self.generate_config: self.max_length = self.generate_config["max_new_tokens"] else: raise ValueError( "either max_tokens (vllm) or max_new_tokens (transformers) must be set in generate_config." ) # Initialize verifiable reward. grpo_config.get("response_format_tags", None) self.global_step = 0 def setup(self): super().setup() if (not self.plugin.pp_size > 1 and self.rank == 0) or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 and self.dp_rank == 0 ): self.wandb_run = wandb.init( project=self.project_name, sync_tensorboard=False, dir="./wandb", name=self.run_name, group=self.wandb_group_name, ) self.lr_scheduler = CosineAnnealingWarmupLR( optimizer=self.optimizer, total_steps=min(self.num_episodes, 4) * self.train_dataset_size // (self.batch_size * self.dp_size), warmup_steps=0, eta_min=0.1 * self.grpo_config.get("lr", 1e-6), ) self.policy_model, self.optimizer, _, _, self.lr_scheduler = self.booster.boost( self.policy_model, self.optimizer, lr_scheduler=self.lr_scheduler ) if self.policy_loss_fn.beta > 0: self.reference_model, *_ = self.booster.boost(self.reference_model) self.plugin.logger.set_level("ERROR") def step(self, pbar: Any, **kwargs) -> Optional[float]: """ Step data from policy model: [{ "input_ids": torch.Tensor, "attention_mask": torch.Tensor, "action_mask": torch.Tensor, "action_log_probs": torch.Tensor, }, ...] Format: [minibatch_size, num_of_generation, prompt_length + response_length] --- <PAD>...<PAD><PROMPT>...<PROMPT><RESPONSE>...<RESPONSE><PAD>...<PAD>. """ # Reshape to [minibatch_size x num_of_generation, prompt_length + response_length] data = {k: v.view(-1, v.size(-1)) for k, v in kwargs.items() if "raw_train_mini_batch_" not in k} self.raw_train_batch_reward.extend(kwargs["raw_train_mini_batch_reward"]) self.raw_train_batch_format_acc.extend(kwargs["raw_train_mini_batch_format_acc"]) self.raw_train_batch_ans_acc.extend(kwargs["raw_train_mini_batch_ans_acc"]) self.raw_train_batch_response_len.extend(kwargs["raw_train_mini_batch_response_len"]) action_mask = data["action_mask"] num_action = action_mask.shape[1] old_action_log_probs = data["action_log_probs"] response_length = torch.sum(action_mask, dim=1).to(torch.float32) train_microbatch_size = self.grpo_config.get("train_microbatch_size", data["input_ids"].size(0)) reward = data["reward"].view((-1)) format_acc = data["format_acc"].view((-1)) ans_acc = data["ans_acc"].view((-1)) # [minibatch_size, num_generations] group_reward = reward.view(-1, self.num_generations) reward_mean = group_reward.mean(dim=1) # [minibatch_size x num_generations] reward_mean = reward_mean.repeat_interleave(self.num_generations, dim=0) reward_std = group_reward.std(dim=1).repeat_interleave(self.num_generations, dim=0) # [minibatch_size x num_generations] advantages = ((reward - reward_mean) / (reward_std + 1e-4)).unsqueeze(dim=-1) # [minibatch_size x num_of_generation] loss_mask = torch.ones(action_mask.size(0), device=action_mask.device).bool() # filter out overlength samples if self.filter_truncated_response and action_mask.size(1) == self.max_length: loss_mask = torch.logical_and( loss_mask, action_mask[:, -1] == False, ) if self.filter_range is not None and self.grpo_config.get("dynamic_batching", False) == False: # filter out samples with reward outside the range # if dynamic batching is enabled, we filter out out of range groups before training group_ans_acc_mean = ( ans_acc.view(-1, self.num_generations).mean(dim=1).repeat_interleave(self.num_generations, dim=-1) ) loss_mask = torch.logical_and( loss_mask, torch.logical_and( group_ans_acc_mean > self.filter_range[0], group_ans_acc_mean < self.filter_range[1], ), ) self.effective_prompt_count += ( group_reward.size(0) * self.dp_size ) # all prompts in the batch are effective as we filtered out the bad ones before step. mean_kl, mean_loss = [], [] need_update = self.effective_prompt_count >= self.batch_size * self.dp_size effective_samples = all_reduce_sum(torch.sum(loss_mask), self.plugin) effective_tokens_count = torch.sum(action_mask, dim=-1) * loss_mask total_effective_tokens_count = all_reduce_sum(torch.sum(effective_tokens_count), self.plugin) self.effective_sample_count += effective_samples.item() pbar.set_postfix( { "Global Step": self.global_step, "Gradient Accumulation on": f"{self.effective_prompt_count}/{self.batch_size * self.dp_size} effective prompts, {self.effective_sample_count}/{self.batch_size * self.dp_size * self.num_generations} effective samples", } ) # Gradient must be synchronized if zero2 is enabled. https://github.com/hpcaitech/ColossalAI/blob/44d4053fec005fe0b06b6bc755fdc962463145df/colossalai/booster/plugin/hybrid_parallel_plugin.py#L1500 ctx = ( nullcontext() if need_update or self.booster.plugin.zero_stage == 2 else self.booster.no_sync(self.policy_model, self.optimizer) ) with ctx: mini_batch_entropies = [] for forward_micro_batch_start in range(0, data["input_ids"].size(0), train_microbatch_size): input_ids_forward_micro_batch = data["input_ids"][ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] old_action_log_probs_micro_batch = old_action_log_probs[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] attention_mask_forward_micro_batch = data["attention_mask"][ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] action_mask_forward_micro_batch = action_mask[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] loss_mask_forward_micro_batch = ( loss_mask[forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size] if loss_mask is not None else None ) advantages_forward_micro_batch = advantages[ forward_micro_batch_start : forward_micro_batch_start + train_microbatch_size ] if self.plugin.pp_size > 1: # Support training with PP. if self.policy_loss_fn.beta > 0: with torch.no_grad(): reference_model_outputs = self.booster.execute_pipeline( iter( [ { "input_ids": input_ids_forward_micro_batch, "attention_mask": attention_mask_forward_micro_batch, } ] ), self.reference_model, criterion=lambda outputs, inputs: torch.tensor( [0.0], device=action_mask.device ), # dummy criterion optimizer=None, return_loss=False, return_outputs=True, ) if self.booster.plugin.stage_manager.is_last_stage(): reference_action_log_probs = memory_efficient_logprob( reference_model_outputs["outputs"]["logits"] / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) else: # Dummy reference logprobs for data iterator. reference_action_log_probs = None else: reference_action_log_probs = None data_policy_forward = { "input_ids": input_ids_forward_micro_batch, "attention_mask": attention_mask_forward_micro_batch, "action_mask": action_mask_forward_micro_batch, "advantages": advantages_forward_micro_batch, "loss_mask": loss_mask_forward_micro_batch, "old_action_log_probs": old_action_log_probs_micro_batch, "source": self.rank, } if reference_action_log_probs is not None: data_policy_forward["reference_action_log_probs"] = reference_action_log_probs kl = [] def _criterion(outputs, inputs): action_logits = outputs.logits mini_batch_entropies.append( ( ((entropy_from_logits(action_logits[:, -num_action:]) * inputs["action_mask"]).sum(-1)) / inputs["action_mask"].sum(-1) ).detach() ) action_log_probs = memory_efficient_logprob( action_logits / self.generate_config["temperature"], inputs["input_ids"], num_action, shard_config=self.plugin.shard_config, ) if "reference_action_log_probs" in inputs: per_token_kl = ( torch.exp(inputs["reference_action_log_probs"] - action_log_probs) - (inputs["reference_action_log_probs"] - action_log_probs) - 1 ) appox_kl = torch.sum(per_token_kl * inputs["action_mask"], dim=-1) / torch.sum( inputs["action_mask"], dim=-1 ) kl.append(appox_kl.mean()) else: per_token_kl = 0.0 kl.append(torch.tensor(0.0)) loss, _ = self.policy_loss_fn( action_log_probs, inputs["old_action_log_probs"], inputs["advantages"].repeat_interleave(action_log_probs.size(-1), dim=-1), per_token_kl, inputs["action_mask"], loss_mask=inputs["loss_mask"], total_effective_tokens_in_batch=total_effective_tokens_count, ) return loss policy_model_outputs = self.booster.execute_pipeline( iter([data_policy_forward]), self.policy_model, criterion=_criterion, optimizer=self.optimizer, return_loss=True, return_outputs=False, ) loss = policy_model_outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): if len(kl) > 0: kl = all_reduce_mean(torch.mean(torch.stack(kl)).to(loss.device), self.plugin).data mean_kl.append(kl) mean_loss.append(all_reduce_mean(loss, self.plugin).data) else: policy_model_logits = self.policy_model( input_ids=input_ids_forward_micro_batch, attention_mask=attention_mask_forward_micro_batch, ).logits action_log_probs = memory_efficient_logprob( policy_model_logits / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) if self.policy_loss_fn.beta > 0: with torch.no_grad(): reference_model_logits = self.reference_model( input_ids=input_ids_forward_micro_batch, attention_mask=attention_mask_forward_micro_batch, ).logits reference_action_log_probs = memory_efficient_logprob( reference_model_logits / self.generate_config["temperature"], input_ids_forward_micro_batch, num_action, shard_config=self.plugin.shard_config, ) per_token_kl = ( torch.exp(reference_action_log_probs - action_log_probs) - (reference_action_log_probs - action_log_probs) - 1 ) kl = torch.sum(per_token_kl * action_mask_forward_micro_batch, dim=-1) / torch.sum( action_mask_forward_micro_batch, dim=-1 ) else: per_token_kl = 0.0 kl = None loss, _ = self.policy_loss_fn( action_log_probs, old_action_log_probs_micro_batch, advantages_forward_micro_batch.repeat_interleave(action_log_probs.size(-1), dim=-1), per_token_kl, action_mask_forward_micro_batch, loss_mask=loss_mask_forward_micro_batch, total_effective_tokens_in_batch=total_effective_tokens_count, ) self.booster.backward(loss, self.optimizer) loss = all_reduce_mean(loss, self.plugin) # Calculate accumulate value. if kl is not None: kl = all_reduce_mean(kl.mean(), self.plugin) mean_kl.append(kl.data) mean_loss.append(loss.data) mini_batch_entropies.append( all_reduce_mean( ( ( ( entropy_from_logits(policy_model_logits[:, -num_action:]) * action_mask_forward_micro_batch ).sum(-1) ) / action_mask_forward_micro_batch.sum(-1) ).detach(), self.plugin, ) ) if not self.plugin.pp_size > 1 or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 and self.dp_rank == 0 ): reward = all_reduce_mean(reward.mean(), self.plugin) format_acc = all_reduce_mean(format_acc.mean(), self.plugin) ans_acc = all_reduce_mean(ans_acc.mean(), self.plugin) advantages = all_reduce_mean(advantages.mean(), self.plugin) response_length = all_reduce_mean(response_length.mean(), self.plugin) entropy = all_reduce_mean(torch.cat(mini_batch_entropies, dim=0).mean(), self.plugin) self.accum_loss.add_(sum(mean_loss) / len(mean_loss)) self.accum_entropy.add_(entropy.data) if self.policy_loss_fn.beta > 0: self.accum_kl.add_(sum(mean_kl) / len(mean_kl)) self.accum_advantages.add_(advantages.data) self.accum_count += 1 if need_update: self.optimizer.step() self.optimizer.zero_grad() self.global_step += 1 if self.lr_scheduler is not None: self.lr_scheduler.step() # no need to run all reduce as raw_train_batch_* are not splited across dp rank sample_utilization = self.effective_sample_count / len(self.raw_train_batch_reward) / self.num_generations self.effective_prompt_count = 0 self.effective_sample_count = 0 loss_scalar = self.accum_loss.item() if not self.plugin.pp_size > 1 or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 ): if (not self.plugin.pp_size > 1 and self.rank == 0) or ( self.plugin.pp_size > 1 and self.booster.plugin.stage_manager.is_last_stage() and self.tp_rank == 0 ): raw_batch_reward_mean = torch.cat(self.raw_train_batch_reward, dim=0).mean().cpu().item() raw_batch_format_acc_mean = torch.cat(self.raw_train_batch_format_acc, dim=0).mean().cpu().item() raw_batch_ans_acc_mean = torch.cat(self.raw_train_batch_ans_acc, dim=0).mean().cpu().item() raw_batch_response_len = torch.cat(self.raw_train_batch_response_len, dim=0) raw_batch_response_len_mean = raw_batch_response_len.mean().cpu().item() overlength_samples_ratio = ( (raw_batch_response_len >= action_mask.size(-1)).to(float).mean().cpu().item() ) # not an exact figure, but a close estimate self.raw_train_batch_reward = [] self.raw_train_batch_format_acc = [] self.raw_train_batch_ans_acc = [] self.raw_train_batch_response_len = [] to_log_msg = [ f"Loss: {self.accum_loss.item() / self.accum_count:.4f}", f"Reward: {raw_batch_reward_mean:.4f}", f"format Reward: {raw_batch_format_acc_mean:.4f}", f"Acc Reward: {raw_batch_ans_acc_mean:.4f}", f"Advantages: {self.accum_advantages.item() / self.accum_count:.4f}", f"Response Length: {raw_batch_response_len_mean:.4f}", f"Sample_utilization: {sample_utilization:.4f}", f"Overlength samples ratio: {overlength_samples_ratio:.4f}", f"Entropy: {self.accum_entropy.item() / self.accum_count:.4f}", ] + ([f"KL: {self.accum_kl.item() / self.accum_count:.4f}"] if self.policy_loss_fn.beta > 0 else []) print("\n".join(to_log_msg)) metrics = { "metrics/reward": raw_batch_reward_mean, "metrics/format_acc": raw_batch_format_acc_mean, "metrics/ans_acc": raw_batch_ans_acc_mean, "metrics/response_length": raw_batch_response_len_mean, "train/loss": self.accum_loss.item() / self.accum_count, "train/advantages": self.accum_advantages.item() / self.accum_count, "train/learning_rate": self.lr_scheduler.get_last_lr()[0], "train/sample_utilization": sample_utilization, "train/entropy": self.accum_entropy.item() / self.accum_count, "train/overlength_samples_ratio": overlength_samples_ratio, "rollout/temperature": data["temperature"].cpu().numpy()[0][0], } if self.policy_loss_fn.beta > 0: metrics["train/kl"] = self.accum_kl.item() / self.accum_count if self.wandb_run is not None: self.wandb_run.log(metrics) ray.get(self.shared_signal_actor.set_signal.remote("sample_utilization", sample_utilization)) self.accum_loss.zero_() self.accum_kl.zero_() self.accum_entropy.zero_() self.accum_advantages.zero_() self.accum_count = 0 return loss_scalar else: return None def state_dict(self): self.policy_model._force_wait_all_gather() model = self.policy_model.unwrap() state_dict = model.state_dict() return state_dict
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/distributed/zero_bubble/__init__.py
applications/ColossalChat/coati/distributed/zero_bubble/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/ppo.py
applications/ColossalChat/coati/trainer/ppo.py
""" PPO trainer """ import os from typing import Dict, List, Optional, Union import torch import wandb from coati.experience_buffer import NaiveExperienceBuffer from coati.experience_maker import Experience, NaiveExperienceMaker from coati.models import Critic, RewardModel, RLVRRewardModel from coati.models.loss import GPTLMLoss, PolicyLoss, ValueLoss from coati.models.utils import calc_action_log_probs from coati.trainer.callbacks import Callback from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader, DistributedSampler from tqdm import tqdm from transformers import PreTrainedModel, PreTrainedTokenizerBase from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import OLTrainer from .utils import CycledDataLoader, is_rank_0, to_device def _set_default_generate_kwargs(actor: PreTrainedModel) -> Dict: """ Set default keyword arguments for generation based on the actor model. Args: actor (PreTrainedModel): The actor model. Returns: Dict: A dictionary containing the default keyword arguments for generation. """ unwrapped_model = actor.unwrap() new_kwargs = {} # use huggingface models method directly if hasattr(unwrapped_model, "prepare_inputs_for_generation"): new_kwargs["prepare_inputs_fn"] = unwrapped_model.prepare_inputs_for_generation if hasattr(unwrapped_model, "_update_model_kwargs_for_generation"): new_kwargs["update_model_kwargs_fn"] = unwrapped_model._update_model_kwargs_for_generation return new_kwargs class PPOTrainer(OLTrainer): """ Trainer for PPO algorithm. Args: strategy (Booster): the strategy to use for training actor (Actor): the actor model in ppo algorithm critic (Critic): the critic model in ppo algorithm reward_model (RewardModel): the reward model in rlhf algorithm to make reward of sentences initial_model (Actor): the initial model in rlhf algorithm to generate reference logics to limit the update of actor actor_optim (Optimizer): the optimizer to use for actor model critic_optim (Optimizer): the optimizer to use for critic model kl_coef (float, defaults to 0.1): the coefficient of kl divergence loss train_batch_size (int, defaults to 8): the batch size to use for training buffer_limit (int, defaults to 0): the max_size limitation of buffer buffer_cpu_offload (bool, defaults to True): whether to offload buffer to cpu eps_clip (float, defaults to 0.2): the clip coefficient of policy loss vf_coef (float, defaults to 1.0): the coefficient of value loss ptx_coef (float, defaults to 0.9): the coefficient of ptx loss value_clip (float, defaults to 0.4): the clip coefficient of value loss sample_buffer (bool, defaults to False): whether to sample from buffer dataloader_pin_memory (bool, defaults to True): whether to pin memory for data loader offload_inference_models (bool, defaults to True): whether to offload inference models to cpu during training process callbacks (List[Callback], defaults to []): the callbacks to call during training process generate_kwargs (dict, optional): the kwargs to use while model generating """ def __init__( self, actor_booster: Booster, critic_booster: Booster, actor: PreTrainedModel, critic: Critic, reward_model: Union[RewardModel, RLVRRewardModel], initial_model: PreTrainedModel, actor_optim: Optimizer, critic_optim: Optimizer, actor_lr_scheduler: _LRScheduler, critic_lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, kl_coef: float = 0.1, ptx_coef: float = 0.9, train_batch_size: int = 8, buffer_limit: int = 0, buffer_cpu_offload: bool = True, eps_clip: float = 0.2, vf_coef: float = 1.0, value_clip: float = 0.2, sample_buffer: bool = False, dataloader_pin_memory: bool = True, offload_inference_models: bool = True, apply_loss_mask: bool = True, accumulation_steps: int = 1, save_interval: int = 0, save_dir: str = None, use_tp: bool = False, coordinator: DistCoordinator = None, callbacks: List[Callback] = [], **generate_kwargs, ) -> None: if isinstance(actor_booster, GeminiPlugin): assert not offload_inference_models, "GeminiPlugin is not compatible with manual model.to('cpu')" data_buffer = NaiveExperienceBuffer(train_batch_size, buffer_limit, buffer_cpu_offload) super().__init__( actor_booster, critic_booster, data_buffer, sample_buffer, dataloader_pin_memory, callbacks=callbacks ) self.generate_kwargs = _set_default_generate_kwargs(actor) self.generate_kwargs.update(generate_kwargs) self.actor = actor self.critic = critic self.actor_booster = actor_booster self.critic_booster = critic_booster self.actor_scheduler = actor_lr_scheduler self.critic_scheduler = critic_lr_scheduler self.tokenizer = tokenizer self.experience_maker = NaiveExperienceMaker( self.actor, self.critic, reward_model, initial_model, self.tokenizer, kl_coef ) self.train_batch_size = train_batch_size self.actor_loss_fn = PolicyLoss(eps_clip) self.critic_loss_fn = ValueLoss(value_clip) self.vf_coef = vf_coef self.ptx_loss_fn = GPTLMLoss() self.ptx_coef = ptx_coef self.actor_optim = actor_optim self.critic_optim = critic_optim self.save_interval = save_interval self.apply_loss_mask = apply_loss_mask self.coordinator = coordinator self.actor_save_dir = os.path.join(save_dir, "actor") self.critic_save_dir = os.path.join(save_dir, "critic") self.num_train_step = 0 self.accumulation_steps = accumulation_steps self.use_tp = use_tp self.accumulative_meter = AccumulativeMeanMeter() self.offload_inference_models = offload_inference_models self.device = get_current_device() def _before_fit( self, prompt_dataloader: DataLoader, pretrain_dataloader: Optional[DataLoader] = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.prompt_dataloader = CycledDataLoader(prompt_dataloader) self.pretrain_dataloader = CycledDataLoader(pretrain_dataloader) if pretrain_dataloader is not None else None self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-ppo", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "ppo") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _setup_update_phrase_dataload(self): """ why not use distributed_dataloader? if tp is used, input on each rank is the same and we use the same dataloader to feed same experience to all ranks if tp is not used, input on each rank is different and we expect different experiences to be fed to each rank """ self.dataloader = DataLoader( self.data_buffer, batch_size=self.train_batch_size, shuffle=True, drop_last=True, pin_memory=self.dataloader_pin_memory, collate_fn=self.data_buffer.collate_fn, ) def _make_experience(self, collect_step: int) -> Experience: """ Make experience """ prompts = self.prompt_dataloader.next() if self.offload_inference_models: # TODO(ver217): this may be controlled by strategy if they are prepared by strategy self.experience_maker.initial_model.to(self.device) self.experience_maker.reward_model.to(self.device) return self.experience_maker.make_experience( input_ids=prompts["input_ids"].to(get_current_device()), attention_mask=prompts["attention_mask"].to(get_current_device()), gt_answer=prompts["gt_answer"], **self.generate_kwargs, ) def _training_step(self, experience: Experience): """ Args: experience: sequences: [batch_size, prompt_length + response_length] --- <PAD>...<PAD><PROMPT>...<PROMPT><RESPONSE>...<RESPONSE><PAD>...<PAD> """ self.actor.train() self.critic.train() num_actions = experience.action_log_probs.size(1) # policy loss actor_logits = self.actor(input_ids=experience.sequences, attention_mask=experience.attention_mask)[ "logits" ] # [batch size, prompt_length + response_length] action_log_probs = calc_action_log_probs(actor_logits, experience.sequences, num_actions) actor_loss, to_skip, max_ratio = self.actor_loss_fn( action_log_probs, experience.action_log_probs, experience.advantages, action_mask=experience.action_mask if self.apply_loss_mask else None, ) actor_loss = (1 - self.ptx_coef) * actor_loss if not to_skip: self.actor_booster.backward(loss=actor_loss, optimizer=self.actor_optim) # ptx loss if self.ptx_coef != 0: batch = self.pretrain_dataloader.next() batch = to_device(batch, self.device) outputs = self.actor(batch["input_ids"], attention_mask=batch["attention_mask"], labels=batch["labels"]) ptx_loss = outputs.loss ptx_loss = self.ptx_coef * ptx_loss self.actor_booster.backward(loss=ptx_loss, optimizer=self.actor_optim) # value loss values = self.critic( input_ids=experience.sequences, attention_mask=experience.attention_mask ) # [batch size, prompt_length + response_length] critic_loss = self.critic_loss_fn( values[:, -num_actions:], experience.values, experience.advantages, action_mask=experience.action_mask if self.apply_loss_mask else None, ) critic_loss = critic_loss * self.vf_coef self.critic_booster.backward(loss=critic_loss, optimizer=self.critic_optim) # sync actor_loss_mean = all_reduce_mean(tensor=actor_loss) critic_loss_mean = all_reduce_mean(tensor=critic_loss) max_ratio_mean = all_reduce_mean(tensor=max_ratio) reward_mean = all_reduce_mean(tensor=experience.reward.mean()) value_mean = all_reduce_mean(tensor=experience.values.mean()) advantages_mean = all_reduce_mean(tensor=experience.advantages.mean()) kl_mean = all_reduce_mean(tensor=experience.kl.mean()) if self.ptx_coef != 0: ptx_loss_mean = all_reduce_mean(tensor=ptx_loss) self.accumulative_meter.add("actor_loss", actor_loss_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("critic_loss", critic_loss_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("max_ratio", max_ratio_mean.to(torch.float16).item()) self.accumulative_meter.add("reward", reward_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("value", value_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("advantages", advantages_mean.to(torch.float16).item()) self.accumulative_meter.add("skip_ratio", 1.0 if to_skip else 0.0) self.accumulative_meter.add("kl", kl_mean.to(torch.float16).item()) if self.ptx_coef != 0: self.accumulative_meter.add("ptx_loss", ptx_loss_mean.to(torch.float16).mean().item()) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: self.actor_optim.step() self.critic_optim.step() self.actor_optim.zero_grad() self.critic_optim.zero_grad() self.actor_scheduler.step() self.critic_scheduler.step() # preparing logging model output and corresponding rewards. if self.num_train_step % 10 == 0: response_text = self.experience_maker.tokenizer.batch_decode( experience.sequences, skip_special_tokens=True ) for i in range(len(response_text)): response_text[i] = response_text[i] + f"\n\nReward: {experience.reward[i]}" if self.writer and is_rank_0() and "wandb_run" in self.__dict__: # log output to wandb my_table = wandb.Table( columns=[f"sample response {i}" for i in range(len(response_text))], data=[response_text] ) try: self.wandb_run.log({"sample_response": my_table}) except OSError as e: self.coordinator.print_on_master(e) elif self.writer and is_rank_0(): for line in response_text: self.coordinator.print_on_master(line) if self.writer and is_rank_0(): self.writer.add_scalar("train/max_ratio", self.accumulative_meter.get("max_ratio"), self.num_train_step) self.writer.add_scalar( "train/skip_ratio", self.accumulative_meter.get("skip_ratio"), self.num_train_step ) self.writer.add_scalar( "train/actor_loss", self.accumulative_meter.get("actor_loss"), self.num_train_step ) self.writer.add_scalar("train/lr_actor", self.actor_optim.param_groups[0]["lr"], self.num_train_step) self.writer.add_scalar("train/lr_critic", self.critic_optim.param_groups[0]["lr"], self.num_train_step) self.writer.add_scalar( "train/critic_loss", self.accumulative_meter.get("critic_loss"), self.num_train_step ) if self.ptx_coef != 0: self.writer.add_scalar( "train/ptx_loss", self.accumulative_meter.get("ptx_loss"), self.num_train_step ) self.writer.add_scalar("reward", self.accumulative_meter.get("reward"), self.num_train_step) self.writer.add_scalar("approx_kl", self.accumulative_meter.get("kl"), self.num_train_step) self.writer.add_scalar("value", self.accumulative_meter.get("value"), self.num_train_step) self.writer.add_scalar("advantages", self.accumulative_meter.get("advantages"), self.num_train_step) self.accumulative_meter.reset() self.num_train_step += 1 def _learn(self, update_step: int): """ Perform the learning step of the PPO algorithm. Args: update_step (int): The current update step. Returns: None """ if self.offload_inference_models: self.experience_maker.initial_model.to("cpu") self.experience_maker.reward_model.to("cpu") # buffer may be empty at first, we should rebuild at each training if self.sample_buffer: experience = self.data_buffer.sample() self._on_learn_batch_start() experience.to_device(self.device) self._training_step(experience) self._on_learn_batch_end(experience) else: if isinstance(self.dataloader.sampler, DistributedSampler): self.dataloader.sampler.set_epoch(update_step) pbar = tqdm(self.dataloader, desc=f"Train epoch [{update_step + 1}]", disable=not is_rank_0()) for experience in pbar: self._on_learn_batch_start() experience.to_device(self.device) self._training_step(experience) self._on_learn_batch_end(experience) def _save_checkpoint(self, episode: int = 0): """ Save the actor and critic checkpoints with running states. Args: episode (int): The current episode number. Returns: None """ self.coordinator.print_on_master("\nStart saving actor checkpoint with running states") save_checkpoint( save_dir=self.actor_save_dir, booster=self.actor_booster, model=self.actor, optimizer=self.actor_optim, lr_scheduler=self.actor_scheduler, epoch=0, step=episode + 1, batch_size=self.train_batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved actor checkpoint at episode {(episode + 1)} at folder {self.actor_save_dir}" ) self.coordinator.print_on_master("\nStart saving critic checkpoint with running states") save_checkpoint( save_dir=self.critic_save_dir, booster=self.critic_booster, model=self.critic, optimizer=self.critic_optim, lr_scheduler=self.critic_scheduler, epoch=0, step=episode + 1, batch_size=self.train_batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved critic checkpoint at episode {(episode + 1)} at folder {self.critic_save_dir}" )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/orpo.py
applications/ColossalChat/coati/trainer/orpo.py
""" Orpo trainer """ import os from typing import Any, Optional import torch from coati.models.loss import OddsRatioLoss from coati.models.utils import calc_masked_log_probs from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader from tqdm import trange from transformers import PreTrainedTokenizerBase from colossalai.booster import Booster, Plugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import SLTrainer from .utils import is_rank_0, to_device class ORPOTrainer(SLTrainer): """ Trainer for ORPO algorithm. Args: actor (Actor): the actor model in ppo algorithm booster (Strategy): the strategy to use for training actor_optim (Optimizer): the optimizer to use for actor model actor_lr_scheduler (_LRScheduler): the lr scheduler to use for actor model tokenizer (PreTrainedTokenizerBase): the tokenizer to use for encoding max_epochs (int, defaults to 1): the max number of epochs to train lam (float, defaults to 0.1): the lambda parameter in ORPO loss accumulation_steps (int): the number of steps to accumulate gradients start_epoch (int, defaults to 0): the start epoch, non-zero if resumed from a checkpoint save_interval (int): the interval to save model checkpoints, default to 0, which means no checkpoint will be saved during trainning save_dir (str): the directory to save checkpoints coordinator (DistCoordinator): the coordinator to use for distributed logging """ def __init__( self, actor: Any, booster: Booster, actor_optim: Optimizer, plugin: Plugin, actor_lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, max_epochs: int = 1, lam: float = 0.1, apply_loss_mask: bool = True, accumulation_steps: int = 1, start_epoch: int = 0, save_interval: int = 0, save_dir: str = None, coordinator: DistCoordinator = None, ) -> None: super().__init__( booster, max_epochs=max_epochs, model=actor, optimizer=actor_optim, plugin=plugin, start_epoch=start_epoch ) self.actor_scheduler = actor_lr_scheduler self.tokenizer = tokenizer self.odds_ratio_loss_fn = OddsRatioLoss() self.save_interval = save_interval self.coordinator = coordinator self.save_dir = save_dir self.num_train_step = 0 self.lam = lam self.apply_loss_mask = apply_loss_mask self.accumulation_steps = accumulation_steps self.device = get_current_device() self.accumulative_meter = AccumulativeMeanMeter() def _before_fit( self, train_preference_dataloader: DataLoader = None, eval_preference_dataloader: DataLoader = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.train_dataloader = train_preference_dataloader self.eval_dataloader = eval_preference_dataloader self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-orpo", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "orpo") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _train(self, epoch: int): """ Args: epoch int: the number of current epoch """ self.model.train() self.accumulative_meter.reset() step_bar = trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) if not self.apply_loss_mask: chosen_loss_mask = chosen_loss_mask.fill_(1.0) reject_loss_mask = reject_loss_mask.fill_(1.0) batch_size = chosen_input_ids.size()[0] actor_out = self.model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), labels=torch.cat( [chosen_input_ids, torch.ones_like(reject_input_ids, dtype=reject_input_ids.dtype) * -100] ), ) torch.autograd.set_detect_anomaly(True) actor_all_logits = actor_out["logits"].to(torch.float32) actor_chosen_logits = actor_all_logits[:batch_size] actor_reject_logits = actor_all_logits[batch_size:] logprob_actor_chosen = calc_masked_log_probs(actor_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:]) logprob_actor_reject = calc_masked_log_probs(actor_reject_logits, reject_input_ids, reject_loss_mask[:, 1:]) # label_chosen[chosen_loss_mask[:, 1:] == 0] = -100 chosen_nll = actor_out["loss"] odds_ratio_loss, log_odds_ratio = self.odds_ratio_loss_fn( logprob_actor_chosen, logprob_actor_reject, chosen_loss_mask[:, 1:], reject_loss_mask[:, 1:] ) loss = chosen_nll - odds_ratio_loss * self.lam step_bar.set_description(f"Epoch {epoch + 1}/{self.max_epochs} Loss: {loss.detach().cpu().item():.4f}") self.booster.backward(loss=loss, optimizer=self.optimizer) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: self.optimizer.step() self.optimizer.zero_grad() self.actor_scheduler.step() chosen_rewards = torch.sum(logprob_actor_chosen) / torch.sum(chosen_loss_mask[:, 1:]) rejected_rewards = torch.sum(logprob_actor_reject) / torch.sum(reject_loss_mask[:, 1:]) reward_accuracies = torch.sum((log_odds_ratio > 0).float()) / torch.sum(log_odds_ratio != 0) # sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_rewards) rejected_rewards_mean = all_reduce_mean(tensor=rejected_rewards) reward_accuracies_mean = all_reduce_mean(tensor=reward_accuracies) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) self.accumulative_meter.add("log_odds_ratio", log_odds_ratio.to(torch.float16).mean().item()) self.accumulative_meter.add("accuracy", reward_accuracies_mean.to(torch.float16).item()) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: step_bar.update() global_step = (self.num_train_step + 1) / self.accumulation_steps # logging if self.writer and is_rank_0(): self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), global_step) self.writer.add_scalar("train/lr", self.optimizer.param_groups[0]["lr"], global_step) self.writer.add_scalar( "train/chosen_rewards", self.accumulative_meter.get("chosen_rewards"), global_step ) self.writer.add_scalar( "train/rejected_rewards", self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/margin", self.accumulative_meter.get("chosen_rewards") - self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/accuracy", self.accumulative_meter.get("accuracy"), global_step, ) self.writer.add_scalar( "train/log_odds_ratio", self.accumulative_meter.get("log_odds_ratio"), global_step, ) self.accumulative_meter.reset() if self.save_dir is not None and (self.num_train_step + 1) % self.save_interval == 0: # save checkpoint self.coordinator.print_on_master("\nStart saving model checkpoint with running states") save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.model, optimizer=self.optimizer, lr_scheduler=self.actor_scheduler, epoch=epoch, step=i + 1, batch_size=batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved checkpoint at epoch {epoch} step {self.save_interval} at folder {self.save_dir}" ) self.num_train_step += 1 step_bar.close() def _eval(self, epoch: int): """ Args: epoch int: the number of current epoch """ if self.eval_dataloader is None: self.coordinator.print_on_master("No eval dataloader is provided, skip evaluation") return self.model.eval() self.coordinator.print_on_master("\nStart evaluation...") step_bar = trange( len(self.eval_dataloader), desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) self.accumulative_meter.reset() with torch.no_grad(): for i, batch in enumerate(self.eval_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) if not self.apply_loss_mask: chosen_loss_mask = chosen_loss_mask.fill_(1.0) reject_loss_mask = reject_loss_mask.fill_(1.0) batch_size = chosen_input_ids.size()[0] actor_out = self.model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), labels=torch.cat( [chosen_input_ids, torch.ones_like(reject_input_ids, dtype=reject_input_ids.dtype) * -100] ), ) torch.autograd.set_detect_anomaly(True) actor_all_logits = actor_out["logits"].to(torch.float32) actor_chosen_logits = actor_all_logits[:batch_size] actor_reject_logits = actor_all_logits[batch_size:] logprob_actor_chosen = calc_masked_log_probs( actor_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:] ) logprob_actor_reject = calc_masked_log_probs( actor_reject_logits, reject_input_ids, reject_loss_mask[:, 1:] ) chosen_nll = actor_out["loss"] odds_ratio_loss, log_odds_ratio = self.odds_ratio_loss_fn( logprob_actor_chosen, logprob_actor_reject, chosen_loss_mask[:, 1:], reject_loss_mask[:, 1:] ) loss = chosen_nll - odds_ratio_loss * self.lam step_bar.set_description(f"Epoch {epoch + 1}/{self.max_epochs} Loss: {loss.detach().cpu().item():.4f}") chosen_rewards = torch.sum(logprob_actor_chosen) / torch.sum(chosen_loss_mask[:, 1:]) rejected_rewards = torch.sum(logprob_actor_reject) / torch.sum(reject_loss_mask[:, 1:]) reward_accuracies = torch.sum((log_odds_ratio > 0).float()) / torch.sum(log_odds_ratio != 0) # sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_rewards) rejected_rewards_mean = all_reduce_mean(tensor=rejected_rewards) reward_accuracies_mean = all_reduce_mean(tensor=reward_accuracies) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) self.accumulative_meter.add("log_odds_ratio", log_odds_ratio.to(torch.float16).mean().item()) self.accumulative_meter.add("accuracy", reward_accuracies_mean.to(torch.float16).item()) msg = "Evaluation Result:\n" for tag in ["loss", "chosen_rewards", "rejected_rewards", "log_odds_ratio", "accuracy"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" self.coordinator.print_on_master(msg) os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/rm.py
applications/ColossalChat/coati/trainer/rm.py
""" Reward model trianer """ import os from typing import Any, Callable, Optional import torch import tqdm from coati.models import LogSigLoss from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader from transformers import PreTrainedTokenizerBase from colossalai.booster import Booster, Plugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import SLTrainer from .utils import is_rank_0, to_device class RewardModelTrainer(SLTrainer): """ Trainer for PPO algorithm. Args: actor (Actor): the actor model in ppo algorithm ref_model (Critic): the reference model in ppo algorithm booster (Strategy): the strategy to use for training actor_optim (Optimizer): the optimizer to use for actor model actor_lr_scheduler (_LRScheduler): the lr scheduler to use for actor model tokenizer (PreTrainedTokenizerBase): the tokenizer to use for encoding max_epochs (int, defaults to 1): the max number of epochs to train beta (float, defaults to 0.1): the beta parameter in dpo loss accumulation_steps (int): the number of steps to accumulate gradients start_epoch (int, defaults to 0): the start epoch, non-zero if resumed from a checkpoint save_interval (int): the interval to save model checkpoints, default to 0, which means no checkpoint will be saved during trainning save_dir (str): the directory to save checkpoints coordinator (DistCoordinator): the coordinator to use for distributed logging """ def __init__( self, model: Any, booster: Booster, optimizer: Optimizer, plugin: Plugin, lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, loss_fn: Optional[Callable] = None, max_epochs: int = 1, beta: float = 0.1, accumulation_steps: int = 1, start_epoch: int = 0, save_interval: int = 0, save_dir: str = None, coordinator: DistCoordinator = None, ) -> None: super().__init__( booster, max_epochs=max_epochs, model=model, optimizer=optimizer, plugin=plugin, start_epoch=start_epoch ) self.actor_scheduler = lr_scheduler self.tokenizer = tokenizer self.loss_fn = loss_fn if loss_fn is not None else LogSigLoss(beta=beta) self.save_interval = save_interval self.coordinator = coordinator self.save_dir = save_dir self.num_train_step = 0 self.accumulation_steps = accumulation_steps self.device = get_current_device() self.accumulative_meter = AccumulativeMeanMeter() def _before_fit( self, train_preference_dataloader: DataLoader = None, eval_preference_dataloader: DataLoader = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.train_dataloader = train_preference_dataloader self.eval_dataloader = eval_preference_dataloader self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-rm", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "rm") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _train(self, epoch): self.model.train() step_bar = tqdm.trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, reject_input_ids, reject_attention_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], ) batch_size = chosen_input_ids.size()[0] # Concatenate for better parrallelism reward = self.model( torch.cat([chosen_input_ids, reject_input_ids], dim=0), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask], dim=0), ) chosen_reward = reward[:batch_size] reject_reward = reward[batch_size:] loss = self.loss_fn(chosen_reward, reject_reward).mean() self.booster.backward(loss=loss, optimizer=self.optimizer) accuracy = (chosen_reward > reject_reward).float() # Sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_reward) rejected_rewards_mean = all_reduce_mean(tensor=reject_reward) accuracy_mean = all_reduce_mean(tensor=accuracy) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) self.accumulative_meter.add("accuracy", accuracy_mean.mean().to(torch.float16).item()) if (self.num_train_step + 1) % self.accumulation_steps == 0: self.optimizer.step() self.optimizer.zero_grad() self.actor_scheduler.step() step_bar.update() # Logging if self.writer and is_rank_0(): global_step = (self.num_train_step + 1) / self.accumulation_steps self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), global_step) self.writer.add_scalar("train/lr", self.optimizer.param_groups[0]["lr"], global_step) self.writer.add_scalar( "train/dist", self.accumulative_meter.get("chosen_rewards") - self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/reward_chosen", self.accumulative_meter.get("chosen_rewards"), global_step ) self.writer.add_scalar( "train/reward_reject", self.accumulative_meter.get("rejected_rewards"), global_step ) self.writer.add_scalar("train/acc", self.accumulative_meter.get("accuracy"), global_step) self.accumulative_meter.reset() # Save checkpoint if self.save_interval > 0 and (self.num_train_step + 1) % self.save_interval == 0: self.coordinator.print_on_master("\nStart saving model checkpoint with running states") save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.model, optimizer=self.optimizer, lr_scheduler=self.actor_scheduler, epoch=epoch, step=i + 1, batch_size=batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved checkpoint at epoch {epoch} step {(i + 1)/self.accumulation_steps} at folder {self.save_dir}" ) self.num_train_step += 1 step_bar.close() def _eval(self, epoch): if self.eval_dataloader is None: self.coordinator.print_on_master("No eval dataloader is provided, skip evaluation") return self.model.eval() step_bar = tqdm.trange( len(self.eval_dataloader), desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0() ) with torch.no_grad(): for i, batch in enumerate(self.eval_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, reject_input_ids, reject_attention_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], ) chosen_reward = self.model(chosen_input_ids, attention_mask=chosen_attention_mask) reject_reward = self.model(reject_input_ids, attention_mask=reject_attention_mask) loss = self.loss_fn(chosen_reward, reject_reward).mean() # Sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_reward) rejected_rewards_mean = all_reduce_mean(tensor=reject_reward) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) step_bar.update() msg = "Evaluation Result:\n" for tag in ["loss", "chosen_rewards", "rejected_rewards"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" msg = ( msg + f"distance: {self.accumulative_meter.get('chosen_rewards')-self.accumulative_meter.get('rejected_rewards')}\n" ) self.coordinator.print_on_master(msg) os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/sft.py
applications/ColossalChat/coati/trainer/sft.py
""" SFT trainer """ import os from typing import Optional import torch import torch.distributed as dist from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader from tqdm import tqdm, trange from colossalai.booster import Booster from colossalai.booster.plugin import HybridParallelPlugin, Plugin from colossalai.cluster import DistCoordinator from .base import SLTrainer from .utils import is_rank_0, to_device class SFTTrainer(SLTrainer): """ Trainer to use while training reward model. Args: model (torch.nn.Module): the model to train strategy (Strategy): the strategy to use for training optim(Optimizer): the optimizer to use for training lr_scheduler(_LRScheduler): the lr scheduler to use for training max_epochs (int, defaults to 2): the number of epochs to train accumulation_steps (int, defaults to 8): the number of steps to accumulate gradients """ def __init__( self, model, booster: Booster, optim: Optimizer, lr_scheduler: _LRScheduler, max_epochs: int = 2, plugin: Plugin = None, accumulation_steps: int = 8, apply_loss_mask: bool = True, start_epoch=0, save_interval: int = None, save_dir: str = None, coordinator: Optional[DistCoordinator] = None, ) -> None: super().__init__(booster, max_epochs, model, optim, plugin, start_epoch=start_epoch) self.accumulation_steps = accumulation_steps self.scheduler = lr_scheduler self.save_interval = save_interval self.save_dir = save_dir self.coordinator = coordinator self.num_train_step = 0 self.num_eval_step = 0 self.apply_loss_mask = apply_loss_mask self.accumulative_meter = AccumulativeMeanMeter() def _before_fit( self, train_dataloader: DataLoader, eval_dataloader: Optional[DataLoader] = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: train_dataloader: the dataloader to use for training eval_dataloader: the dataloader to use for evaluation log_dir: the directory to save logs use_wandb: whether to use wandb for logging """ self.train_dataloader = train_dataloader self.eval_dataloader = eval_dataloader self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb wandb.init(project="Coati-sft", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "sft") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _train(self, epoch: int): self.model.train() if isinstance(self.plugin, HybridParallelPlugin) and self.plugin.pp_size > 1: data_iter = iter(self.train_dataloader) step_bar = tqdm( range(len(self.train_dataloader)), desc="Step", disable=not (dist.get_rank() == dist.get_world_size() - 1), ) for step in step_bar: outputs = self.booster.execute_pipeline( data_iter, self.model, criterion=lambda outputs, inputs: outputs[0], optimizer=self.optimizer, return_loss=True, ) loss = outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): global_loss = all_reduce_mean(loss, self.plugin) if dist.get_rank() == dist.get_world_size() - 1: step_bar.set_postfix({"train/loss": global_loss.item()}) self.optimizer.step() self.optimizer.zero_grad() else: step_bar = trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, torch.cuda.current_device()) batch_size = batch["input_ids"].size(0) outputs = self.model( batch["input_ids"], attention_mask=batch["attention_mask"], labels=batch["labels"] if self.apply_loss_mask else batch["input_ids"], ) loss = outputs.loss self.booster.backward(loss=loss, optimizer=self.optimizer) loss_mean = all_reduce_mean(tensor=loss) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) # Gradient accumulation if (self.num_train_step + 1) % self.accumulation_steps == 0: self.optimizer.step() self.optimizer.zero_grad() self.scheduler.step() global_step = (self.num_train_step + 1) / self.accumulation_steps step_bar.set_postfix({"train/loss": self.accumulative_meter.get("loss")}) if self.writer: self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), global_step) self.writer.add_scalar("train/lr", self.scheduler.get_last_lr()[0], global_step) self.accumulative_meter.reset() step_bar.update() self.num_train_step += 1 # Save checkpoint if ( self.save_dir is not None and self.save_interval is not None and (self.num_train_step + 1) % self.save_interval == 0 ): save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.model, optimizer=self.optimizer, lr_scheduler=self.scheduler, epoch=epoch, step=self.num_train_step + 1, batch_size=batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved checkpoint at epoch {epoch} step {self.num_train_step} at folder {self.save_dir}" ) step_bar.close() def _eval(self, epoch: int): if self.eval_dataloader is None: self.coordinator.print_on_master("No eval dataloader is provided, skip evaluation") return self.accumulative_meter.reset() self.model.eval() with torch.no_grad(): if isinstance(self.plugin, HybridParallelPlugin) and self.plugin.pp_size > 1: data_iter = iter(self.eval_dataloader) step_bar = tqdm( range(len(self.eval_dataloader)), desc="Step", disable=not (dist.get_rank() == dist.get_world_size() - 1), ) for step in step_bar: outputs = self.booster.execute_pipeline( data_iter, self.model, criterion=lambda outputs, inputs: outputs[0], optimizer=self.optimizer, return_loss=True, ) loss = outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): global_loss = all_reduce_mean(loss, self.plugin) if dist.get_rank() == dist.get_world_size() - 1: step_bar.set_postfix({"eval/loss": global_loss.item()}) self.accumulative_meter.add("loss", global_loss.item()) if dist.get_rank() == dist.get_world_size() - 1: loss_mean = self.accumulative_meter.get("loss") msg = "Evaluation Result:\n" for tag in ["loss"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" print(msg) if self.save_dir is not None: os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close() else: step_bar = trange( len(self.eval_dataloader), desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for batch in self.eval_dataloader: batch = to_device(batch, torch.cuda.current_device()) outputs = self.model( batch["input_ids"], attention_mask=batch["attention_mask"], labels=batch["labels"] if self.apply_loss_mask else batch["input_ids"], ) loss_mean = all_reduce_mean(tensor=outputs.loss) self.accumulative_meter.add("loss", loss_mean.item(), count_update=batch["input_ids"].size(0)) step_bar.update() loss_mean = self.accumulative_meter.get("loss") msg = "Evaluation Result:\n" for tag in ["loss"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" self.coordinator.print_on_master(msg) if self.save_dir is not None: os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/utils.py
applications/ColossalChat/coati/trainer/utils.py
""" Training utilities for Coati. """ from typing import Any import torch import torch.distributed as dist from torch.utils._pytree import tree_map from torch.utils.data import DataLoader from colossalai.booster import Plugin class AnnealingScheduler: def __init__(self, start, end, warmup_steps=100, annealing_step=2000): self.start = start self.end = end self.warmup_steps = warmup_steps self.step = 0 self.annealing_step = annealing_step def get_temperature(self): if self.step <= self.warmup_steps: return self.start # Stop annealing after warm-up steps elif self.step >= self.annealing_step: return self.end # Linear annealing temp = self.start - (self.step / self.annealing_step) * (self.start - self.end) return temp def step_forward(self): self.step += 1 class CycledDataLoader: """ A data loader that cycles through the data when it reaches the end. Args: dataloader (DataLoader): The original data loader. Attributes: dataloader (DataLoader): The original data loader. count (int): The number of times the data loader has been cycled. dataloader_iter (iterable): The iterator for the data loader. Methods: next(): Returns the next batch of data from the data loader, cycling through the data if necessary. """ def __init__( self, dataloader: DataLoader, ) -> None: self.dataloader = dataloader self.count = 0 self.dataloader_iter = None def next(self): """ Returns the next batch of data from the data loader, cycling through the data if necessary. Returns: Any: The next batch of data from the data loader. """ # defer initialization if self.dataloader_iter is None: self.dataloader_iter = iter(self.dataloader) self.count += 1 try: return next(self.dataloader_iter) except StopIteration: self.count = 0 self.dataloader_iter = iter(self.dataloader) return next(self.dataloader_iter) def is_rank_0() -> bool: """ Check if the current process is the rank 0 process in a distributed training setup. Returns: bool: True if the current process is the rank 0 process, False otherwise. """ return not dist.is_initialized() or dist.get_rank() == 0 def to_device(x: Any, device: torch.device) -> Any: """ Move the input tensor or nested structure of tensors to the specified device. Args: x (Any): The input tensor or nested structure of tensors. device (torch.device): The target device to move the tensors to. Returns: Any: The tensor or nested structure of tensors moved to the target device. """ def _to(t: Any): if isinstance(t, torch.Tensor): return t.to(device) return t return tree_map(_to, x) def all_reduce_mean(tensor: torch.Tensor, plugin: Plugin = None) -> torch.Tensor: """ Perform all-reduce operation on the given tensor and compute the mean across all processes. Args: tensor (torch.Tensor): The input tensor to be reduced. Returns: torch.Tensor: The reduced tensor with mean computed across all processes. """ # All reduce mean across DP group if plugin is not None: dist.all_reduce(tensor=tensor, op=dist.ReduceOp.SUM, group=plugin.dp_group) tensor.div_(plugin.dp_size) else: dist.all_reduce(tensor=tensor, op=dist.ReduceOp.SUM) tensor.div_(dist.get_world_size()) return tensor def all_reduce_sum(tensor: torch.Tensor, plugin: Plugin = None) -> torch.Tensor: """ Performs an all-reduce operation to sum the values of the given tensor across all processes. Args: tensor (torch.Tensor): The input tensor to be reduced. Returns: torch.Tensor: The reduced tensor with the sum of values across all processes. """ # All reduce sum across DP group if plugin is not None: dist.all_reduce(tensor=tensor, op=dist.ReduceOp.SUM, group=plugin.dp_group) else: dist.all_reduce(tensor=tensor, op=dist.ReduceOp.SUM) return tensor def all_gather_tensors(local_tensor_list: torch.Tensor, plugin: Plugin = None) -> torch.Tensor: """ Gathers tensors from all processes and concatenates them along the first dimension. Args: tensor (torch.Tensor): The input tensor to be gathered. Returns: torch.Tensor: The gathered tensor. """ # Gather tensors across DP group if plugin is not None: all_tensor_lists = [None] * plugin.dp_size dist.all_gather_object(all_tensor_lists, local_tensor_list, group=plugin.dp_group) gathered_tensor_list = [] for tensors in all_tensor_lists: gathered_tensor_list.extend(tensors) else: all_tensor_lists = [None] * dist.get_world_size() dist.all_gather_object(all_tensor_lists, local_tensor_list) gathered_tensor_list = [] for tensors in all_tensor_lists: gathered_tensor_list.extend(tensors) return gathered_tensor_list
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/kto.py
applications/ColossalChat/coati/trainer/kto.py
""" KTO trainer """ import os from typing import Any, Optional import torch import torch.distributed as dist from coati.models.loss import KTOLoss from coati.models.utils import calc_masked_log_probs from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader from tqdm import trange from transformers import PreTrainedTokenizerBase from colossalai.booster import Booster, Plugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import SLTrainer from .utils import is_rank_0, to_device class KTOTrainer(SLTrainer): """ Trainer for KTO algorithm. Args: actor (Actor): the actor model in ppo algorithm ref_model (Critic): the reference model in ppo algorithm booster (Strategy): the strategy to use for training actor_optim (Optimizer): the optimizer to use for actor model actor_lr_scheduler (_LRScheduler): the lr scheduler to use for actor model tokenizer (PreTrainedTokenizerBase): the tokenizer to use for encoding max_epochs (int, defaults to 1): the max number of epochs to train accumulation_steps (int): the number of steps to accumulate gradients start_epoch (int, defaults to 0): the start epoch, non-zero if resumed from a checkpoint save_interval (int): the interval to save model checkpoints, default to 0, which means no checkpoint will be saved during trainning save_dir (str): the directory to save checkpoints coordinator (DistCoordinator): the coordinator to use for distributed logging beta (float, defaults to 0.1): the beta parameter in kto loss desirable_weight (float, defaults to 1.0): the weight for desirable reward undesirable_weight (float, defaults to 1.0): the weight for undesirable reward """ def __init__( self, actor: Any, ref_model: Any, booster: Booster, actor_optim: Optimizer, plugin: Plugin, actor_lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, max_epochs: int = 1, beta: float = 0.1, desirable_weight: float = 1.0, undesirable_weight: float = 1.0, apply_loss_mask: bool = True, accumulation_steps: int = 1, start_epoch: int = 0, save_interval: int = 0, save_dir: str = None, coordinator: DistCoordinator = None, ) -> None: super().__init__( booster, max_epochs=max_epochs, model=actor, optimizer=actor_optim, plugin=plugin, start_epoch=start_epoch ) self.ref_model = ref_model self.actor_scheduler = actor_lr_scheduler self.tokenizer = tokenizer self.kto_loss = KTOLoss(beta=beta, desirable_weight=desirable_weight, undesirable_weight=undesirable_weight) self.apply_loss_mask = apply_loss_mask self.save_interval = save_interval self.coordinator = coordinator self.save_dir = save_dir self.num_train_step = 0 self.accumulation_steps = accumulation_steps self.device = get_current_device() self.accumulative_meter = AccumulativeMeanMeter() self.desirable_weight = desirable_weight self.undesirable_weight = undesirable_weight self.beta = beta def _before_fit( self, train_preference_dataloader: DataLoader = None, eval_preference_dataloader: DataLoader = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.train_dataloader = train_preference_dataloader self.eval_dataloader = eval_preference_dataloader self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-kto", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "kto") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _train(self, epoch: int): """ Args: epoch int: the number of current epoch """ self.model.train() self.accumulative_meter.reset() step_bar = trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) (input_ids, attention_mask, loss_mask, label, kl_input_ids, kl_attention_mask, kl_loss_mask) = ( batch["input_ids"], batch["attention_mask"], batch["loss_mask"], batch["label"], batch["kl_input_ids"], batch["kl_attention_mask"], batch["kl_loss_mask"], ) if not self.apply_loss_mask: loss_mask = loss_mask.fill_(1.0) kl_loss_mask = kl_loss_mask.fill_(1.0) batch_size = input_ids.size()[0] # actor logits with torch.no_grad(): # calculate KL term with KT data kl_logits = self.model( input_ids=kl_input_ids, attention_mask=kl_attention_mask, )["logits"] logits = self.model( input_ids=input_ids, attention_mask=attention_mask, )["logits"] logprob = calc_masked_log_probs(logits, input_ids, loss_mask[:, 1:]).sum(-1) kl_logprob = calc_masked_log_probs(kl_logits, kl_input_ids, kl_loss_mask[:, 1:]).sum(-1) chosen_index = [i for i in range(batch_size) if label[i] == 1] rejected_index = [i for i in range(batch_size) if label[i] == 0] chosen_logprob = logprob[chosen_index] rejected_logprob = logprob[rejected_index] with torch.no_grad(): ref_kl_logits = self.ref_model( input_ids=kl_input_ids, attention_mask=kl_attention_mask, )["logits"] ref_logits = self.ref_model( input_ids=input_ids, attention_mask=attention_mask, )["logits"] ref_logprob = calc_masked_log_probs(ref_logits, input_ids, loss_mask[:, 1:]).sum(-1) ref_kl_logprob = calc_masked_log_probs(ref_kl_logits, kl_input_ids, kl_loss_mask[:, 1:]).sum(-1) ref_chosen_logprob = ref_logprob[chosen_index] ref_rejected_logprob = ref_logprob[rejected_index] loss, chosen_rewards, rejected_rewards, kl = self.kto_loss( chosen_logprob, rejected_logprob, kl_logprob, ref_chosen_logprob, ref_rejected_logprob, ref_kl_logprob ) self.booster.backward(loss=loss, optimizer=self.optimizer) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: self.optimizer.step() self.optimizer.zero_grad() self.actor_scheduler.step() # sync loss_mean = all_reduce_mean(tensor=loss) chosen_reward_mean = chosen_rewards.mean() chosen_rewards_list = [ torch.tensor(0, dtype=chosen_reward_mean.dtype, device=loss.device) for _ in range(dist.get_world_size()) ] dist.all_gather(chosen_rewards_list, chosen_reward_mean) rejected_reward_mean = rejected_rewards.mean() rejected_rewards_list = [ torch.tensor(0, dtype=rejected_reward_mean.dtype, device=loss.device) for _ in range(dist.get_world_size()) ] dist.all_gather(rejected_rewards_list, rejected_reward_mean) chosen_rewards_list = [i for i in chosen_rewards_list if not i.isnan()] rejected_rewards_list = [i for i in rejected_rewards_list if not i.isnan()] chosen_rewards_mean = ( torch.stack(chosen_rewards_list).mean() if len(chosen_rewards_list) > 0 else torch.tensor(torch.nan, dtype=loss.dtype, device=loss.device) ) rejected_rewards_mean = ( torch.stack(rejected_rewards_list).mean() if len(rejected_rewards_list) > 0 else torch.tensor(torch.nan, dtype=loss.dtype, device=loss.device) ) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).detach().item()) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: step_bar.update() # logging if self.writer and is_rank_0(): global_step = (self.num_train_step + 1) / self.accumulation_steps self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), global_step) self.writer.add_scalar("train/lr", self.optimizer.param_groups[0]["lr"], global_step) self.writer.add_scalar( "train/chosen_rewards", self.accumulative_meter.get("chosen_rewards"), global_step ) self.writer.add_scalar( "train/rejected_rewards", self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/margin", self.accumulative_meter.get("chosen_rewards") - self.accumulative_meter.get("rejected_rewards"), global_step, ) self.accumulative_meter.reset() if self.save_dir is not None and (self.num_train_step + 1) % self.save_interval == 0: # save checkpoint self.coordinator.print_on_master("\nStart saving model checkpoint with running states") save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.model, optimizer=self.optimizer, lr_scheduler=self.actor_scheduler, epoch=epoch, step=i + 1, batch_size=batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved checkpoint at epoch {epoch} step {self.save_interval} at folder {self.save_dir}" ) self.num_train_step += 1 step_bar.close() def _eval(self, epoch: int): """ Args: epoch int: the number of current epoch """ if self.eval_dataloader is None: self.coordinator.print_on_master("No eval dataloader is provided, skip evaluation") return self.model.eval() self.accumulative_meter.reset() step_bar = trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) (input_ids, attention_mask, loss_mask, label, kl_input_ids, kl_attention_mask, kl_loss_mask) = ( batch["input_ids"], batch["attention_mask"], batch["loss_mask"], batch["label"], batch["kl_input_ids"], batch["kl_attention_mask"], batch["kl_loss_mask"], ) if not self.apply_loss_mask: loss_mask = loss_mask.fill_(1.0) kl_loss_mask = kl_loss_mask.fill_(1.0) batch_size = input_ids.size()[0] # actor logits with torch.no_grad(): # calculate KL term with KT data kl_logits = self.model( input_ids=kl_input_ids, attention_mask=kl_attention_mask, )["logits"] logits = self.model( input_ids=input_ids, attention_mask=attention_mask, )["logits"] logprob = calc_masked_log_probs(logits, input_ids, loss_mask[:, 1:]).sum(-1) kl_logprob = calc_masked_log_probs(kl_logits, kl_input_ids, kl_loss_mask[:, 1:]).sum(-1) chosen_index = [i for i in range(batch_size) if label[i] == 1] rejected_index = [i for i in range(batch_size) if label[i] == 0] chosen_logprob = logprob[chosen_index] rejected_logprob = logprob[rejected_index] with torch.no_grad(): ref_kl_logits = self.ref_model( input_ids=kl_input_ids, attention_mask=kl_attention_mask, )["logits"] ref_logits = self.ref_model( input_ids=input_ids, attention_mask=attention_mask, )["logits"] ref_logprob = calc_masked_log_probs(ref_logits, input_ids, loss_mask[:, 1:]).sum(-1) ref_kl_logprob = calc_masked_log_probs(ref_kl_logits, kl_input_ids, kl_loss_mask[:, 1:]).sum(-1) ref_chosen_logprob = ref_logprob[chosen_index] ref_rejected_logprob = ref_logprob[rejected_index] loss, chosen_rewards, rejected_rewards, kl = self.kto_loss( chosen_logprob, rejected_logprob, kl_logprob, ref_chosen_logprob, ref_rejected_logprob, ref_kl_logprob ) # sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_rewards.mean()) rejected_rewards_mean = all_reduce_mean(tensor=rejected_rewards.mean()) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).detach().item()) self.accumulative_meter.add( "margin", (chosen_rewards_mean - rejected_rewards_mean).to(torch.float16).mean().item() ) step_bar.update() msg = "Evaluation Result:\n" for tag in ["loss", "chosen_rewards", "rejected_rewards", "margin"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" self.coordinator.print_on_master(msg) os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/__init__.py
applications/ColossalChat/coati/trainer/__init__.py
from .base import OLTrainer, SLTrainer from .dpo import DPOTrainer from .grpo import GRPOTrainer from .kto import KTOTrainer from .orpo import ORPOTrainer from .ppo import PPOTrainer from .rm import RewardModelTrainer from .sft import SFTTrainer __all__ = [ "SLTrainer", "OLTrainer", "RewardModelTrainer", "SFTTrainer", "PPOTrainer", "DPOTrainer", "ORPOTrainer", "KTOTrainer", "GRPOTrainer", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/grpo.py
applications/ColossalChat/coati/trainer/grpo.py
""" GRPO trainer """ import os from typing import Dict, List, Optional, Union import torch import wandb from coati.experience_buffer import NaiveExperienceBuffer from coati.experience_maker import Experience, NaiveExperienceMaker from coati.models import RewardModel, RLVRRewardModel from coati.models.loss import GPTLMLoss, PolicyLoss from coati.models.utils import calc_action_log_probs from coati.trainer.callbacks import Callback from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader, DistributedSampler from tqdm import tqdm from transformers import PreTrainedModel, PreTrainedTokenizerBase from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import OLTrainer from .utils import AnnealingScheduler, CycledDataLoader, is_rank_0, to_device def _set_default_generate_kwargs(actor: PreTrainedModel) -> Dict: """ Set default keyword arguments for generation based on the actor model. Args: actor (PreTrainedModel): The actor model. Returns: Dict: A dictionary containing the default keyword arguments for generation. """ unwrapped_model = actor.unwrap() new_kwargs = {} # use huggingface models method directly if hasattr(unwrapped_model, "prepare_inputs_for_generation"): new_kwargs["prepare_inputs_fn"] = unwrapped_model.prepare_inputs_for_generation if hasattr(unwrapped_model, "_update_model_kwargs_for_generation"): new_kwargs["update_model_kwargs_fn"] = unwrapped_model._update_model_kwargs_for_generation return new_kwargs class GRPOTrainer(OLTrainer): """ Trainer for GRPO algorithm. Args: strategy (Booster): the strategy to use for training actor (Actor): the actor model in ppo algorithm reward_model (RewardModel): the reward model in rlhf algorithm to make reward of sentences initial_model (Actor): the initial model in rlhf algorithm to generate reference logics to limit the update of actor actor_optim (Optimizer): the optimizer to use for actor model kl_coef (float, defaults to 0.1): the coefficient of kl divergence loss train_batch_size (int, defaults to 8): the batch size to use for training buffer_limit (int, defaults to 0): the max_size limitation of buffer buffer_cpu_offload (bool, defaults to True): whether to offload buffer to cpu eps_clip (float, defaults to 0.2): the clip coefficient of policy loss vf_coef (float, defaults to 1.0): the coefficient of value loss ptx_coef (float, defaults to 0.9): the coefficient of ptx loss value_clip (float, defaults to 0.4): the clip coefficient of value loss sample_buffer (bool, defaults to False): whether to sample from buffer dataloader_pin_memory (bool, defaults to True): whether to pin memory for data loader offload_inference_models (bool, defaults to True): whether to offload inference models to cpu during training process callbacks (List[Callback], defaults to []): the callbacks to call during training process generate_kwargs (dict, optional): the kwargs to use while model generating """ def __init__( self, actor_booster: Booster, actor: PreTrainedModel, reward_model: Union[RewardModel, RLVRRewardModel], initial_model: PreTrainedModel, actor_optim: Optimizer, actor_lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, kl_coef: float = 0.1, ptx_coef: float = 0.9, train_batch_size: int = 8, buffer_limit: int = 0, buffer_cpu_offload: bool = True, eps_clip: float = 0.2, vf_coef: float = 1.0, value_clip: float = 0.2, sample_buffer: bool = False, dataloader_pin_memory: bool = True, offload_inference_models: bool = True, apply_loss_mask: bool = True, accumulation_steps: int = 1, save_interval: int = 0, save_dir: str = None, use_tp: bool = False, num_generation: int = 8, inference_batch_size: int = None, logits_forward_batch_size: int = None, temperature_annealing_config: Optional[Dict] = None, coordinator: DistCoordinator = None, callbacks: List[Callback] = [], **generate_kwargs, ) -> None: if isinstance(actor_booster, GeminiPlugin): assert not offload_inference_models, "GeminiPlugin is not compatible with manual model.to('cpu')" data_buffer = NaiveExperienceBuffer(train_batch_size, buffer_limit, buffer_cpu_offload) super().__init__(actor_booster, None, data_buffer, sample_buffer, dataloader_pin_memory, callbacks=callbacks) self.generate_kwargs = _set_default_generate_kwargs(actor) self.generate_kwargs.update(generate_kwargs) self.actor = actor self.actor_booster = actor_booster self.actor_scheduler = actor_lr_scheduler self.tokenizer = tokenizer self.experience_maker = NaiveExperienceMaker( self.actor, None, reward_model, initial_model, self.tokenizer, kl_coef, use_grpo=True, num_generation=num_generation, inference_batch_size=inference_batch_size, logits_forward_batch_size=logits_forward_batch_size, ) if temperature_annealing_config: # use annealing self.temperature_annealing_scheduler = AnnealingScheduler( temperature_annealing_config["start_temperature"], temperature_annealing_config["end_temperature"], temperature_annealing_config["annealing_warmup_steps"], temperature_annealing_config["annealing_steps"], ) else: self.temperature_annealing_scheduler = None self.train_batch_size = train_batch_size self.actor_loss_fn = PolicyLoss(eps_clip) self.vf_coef = vf_coef self.ptx_loss_fn = GPTLMLoss() self.ptx_coef = ptx_coef self.actor_optim = actor_optim self.save_interval = save_interval self.apply_loss_mask = apply_loss_mask self.coordinator = coordinator self.actor_save_dir = os.path.join(save_dir, "actor") self.num_train_step = 0 self.accumulation_steps = accumulation_steps self.use_tp = use_tp self.accumulative_meter = AccumulativeMeanMeter() self.offload_inference_models = offload_inference_models self.device = get_current_device() def _before_fit( self, prompt_dataloader: DataLoader, pretrain_dataloader: Optional[DataLoader] = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.prompt_dataloader = CycledDataLoader(prompt_dataloader) self.pretrain_dataloader = CycledDataLoader(pretrain_dataloader) if pretrain_dataloader is not None else None self.writer = None if use_wandb and is_rank_0(): assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-grpo", sync_tensorboard=True) if log_dir is not None and is_rank_0(): import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "grpo") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _setup_update_phrase_dataload(self): """ why not use distributed_dataloader? if tp is used, input on each rank is the same and we use the same dataloader to feed same experience to all ranks if tp is not used, input on each rank is different and we expect different experiences to be fed to each rank """ self.dataloader = DataLoader( self.data_buffer, batch_size=self.train_batch_size, shuffle=True, drop_last=True, pin_memory=self.dataloader_pin_memory, collate_fn=self.data_buffer.collate_fn, ) def _make_experience(self, collect_step: int) -> Experience: """ Make experience """ prompts = self.prompt_dataloader.next() if self.offload_inference_models: # TODO(ver217): this may be controlled by strategy if they are prepared by strategy self.experience_maker.initial_model.to(self.device) self.experience_maker.reward_model.to(self.device) if self.temperature_annealing_scheduler: self.generate_kwargs["temperature"] = self.temperature_annealing_scheduler.get_temperature() return self.experience_maker.make_experience( input_ids=prompts["input_ids"].to(get_current_device()), attention_mask=prompts["attention_mask"].to(get_current_device()), gt_answer=prompts["gt_answer"], **self.generate_kwargs, ) def _training_step(self, experience: Experience): """ Args: experience: sequences: [batch_size, prompt_length + response_length] --- <PAD>...<PAD><PROMPT>...<PROMPT><RESPONSE>...<RESPONSE><PAD>...<PAD> """ self.actor.train() num_actions = experience.action_log_probs.size(1) # policy loss actor_logits = self.actor(input_ids=experience.sequences, attention_mask=experience.attention_mask)[ "logits" ] # [batch size, prompt_length + response_length] action_log_probs = calc_action_log_probs(actor_logits, experience.sequences, num_actions) actor_loss, to_skip, max_ratio = self.actor_loss_fn( action_log_probs, experience.action_log_probs, experience.advantages.unsqueeze(dim=-1).repeat_interleave(action_log_probs.size(-1), dim=-1), action_mask=experience.action_mask if self.apply_loss_mask else None, ) # sequence that is not end properly are not counted in token cost token_cost = torch.sum( (experience.sequences[:, -num_actions:] != self.tokenizer.pad_token_id).to(torch.float), axis=-1 ).to(actor_logits.device) end_properly = experience.sequences[:, -1] == self.tokenizer.pad_token_id mean_token_cost = torch.sum(token_cost * end_properly) / torch.sum(end_properly) actor_loss = (1 - self.ptx_coef) * actor_loss if not to_skip: self.actor_booster.backward(loss=actor_loss, optimizer=self.actor_optim) # ptx loss if self.ptx_coef != 0: batch = self.pretrain_dataloader.next() batch = to_device(batch, self.device) outputs = self.actor(batch["input_ids"], attention_mask=batch["attention_mask"], labels=batch["labels"]) ptx_loss = outputs.loss ptx_loss = self.ptx_coef * ptx_loss self.actor_booster.backward(loss=ptx_loss, optimizer=self.actor_optim) # sync actor_loss_mean = all_reduce_mean(tensor=actor_loss) max_ratio_mean = all_reduce_mean(tensor=max_ratio) reward_mean = all_reduce_mean(tensor=experience.reward.mean()) advantages_mean = all_reduce_mean(tensor=experience.advantages.mean()) kl_mean = all_reduce_mean(tensor=experience.kl.mean()) mean_token_cost = all_reduce_mean(tensor=mean_token_cost) if self.ptx_coef != 0: ptx_loss_mean = all_reduce_mean(tensor=ptx_loss) self.accumulative_meter.add("actor_loss", actor_loss_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("max_ratio", max_ratio_mean.to(torch.float16).item()) self.accumulative_meter.add("reward", reward_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("advantages", advantages_mean.to(torch.float16).item()) self.accumulative_meter.add("skip_ratio", 1.0 if to_skip else 0.0) self.accumulative_meter.add("mean_token_cost", mean_token_cost.to(torch.float16).item()) self.accumulative_meter.add("kl", kl_mean.to(torch.float16).item()) if self.ptx_coef != 0: self.accumulative_meter.add("ptx_loss", ptx_loss_mean.to(torch.float16).mean().item()) if self.num_train_step % self.accumulation_steps == self.accumulation_steps - 1: self.actor_optim.step() self.actor_optim.zero_grad() self.actor_scheduler.step() if self.temperature_annealing_scheduler: self.temperature_annealing_scheduler.step_forward() # preparing logging model output and corresponding rewards. if self.num_train_step % 10 == 0: response_text = self.experience_maker.tokenizer.batch_decode( experience.sequences, skip_special_tokens=True ) for i in range(len(response_text)): response_text[i] = response_text[i] + f"\n\nReward: {experience.reward[i]}" if self.writer and is_rank_0() and "wandb_run" in self.__dict__: # log output to wandb my_table = wandb.Table( columns=[f"sample response {i}" for i in range(len(response_text))], data=[response_text] ) try: self.wandb_run.log({"sample_response": my_table}) except OSError as e: self.coordinator.print_on_master(e) elif self.writer and is_rank_0(): for line in response_text: self.coordinator.print_on_master(line) if self.writer and is_rank_0(): global_step = (self.num_train_step + 1) / self.accumulation_steps self.writer.add_scalar("train/max_ratio", self.accumulative_meter.get("max_ratio"), global_step) self.writer.add_scalar("train/skip_ratio", self.accumulative_meter.get("skip_ratio"), global_step) self.writer.add_scalar("train/actor_loss", self.accumulative_meter.get("actor_loss"), global_step) self.writer.add_scalar("train/lr_actor", self.actor_optim.param_groups[0]["lr"], global_step) if self.ptx_coef != 0: self.writer.add_scalar("train/ptx_loss", self.accumulative_meter.get("ptx_loss"), global_step) self.writer.add_scalar("reward", self.accumulative_meter.get("reward"), global_step) self.writer.add_scalar("token_cost", self.accumulative_meter.get("mean_token_cost"), global_step) self.writer.add_scalar("approx_kl", self.accumulative_meter.get("kl"), global_step) self.writer.add_scalar("advantages", self.accumulative_meter.get("advantages"), global_step) self.accumulative_meter.reset() self.num_train_step += 1 def _learn(self, update_step: int): """ Perform the learning step of the PPO algorithm. Args: update_step (int): The current update step. Returns: None """ if self.offload_inference_models: self.experience_maker.initial_model.to("cpu") self.experience_maker.reward_model.to("cpu") # buffer may be empty at first, we should rebuild at each training if self.sample_buffer: experience = self.data_buffer.sample() self._on_learn_batch_start() experience.to_device(self.device) self._training_step(experience) self._on_learn_batch_end(experience) else: if isinstance(self.dataloader.sampler, DistributedSampler): self.dataloader.sampler.set_epoch(update_step) pbar = tqdm(self.dataloader, desc=f"Train epoch [{update_step + 1}]", disable=not is_rank_0()) for experience in pbar: self._on_learn_batch_start() experience.to_device(self.device) self._training_step(experience) self._on_learn_batch_end(experience) def _save_checkpoint(self, num_train_step: int = 0): """ Save the actor checkpoints with running states. Args: num_train_step (int): The current num_train_step number. Returns: None """ self.coordinator.print_on_master("\nStart saving actor checkpoint with running states") save_checkpoint( save_dir=self.actor_save_dir, booster=self.actor_booster, model=self.actor, optimizer=self.actor_optim, lr_scheduler=self.actor_scheduler, epoch=0, step=num_train_step + 1, batch_size=self.train_batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved actor checkpoint at episode {(num_train_step + 1)} at folder {self.actor_save_dir}" )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/base.py
applications/ColossalChat/coati/trainer/base.py
""" Base trainers for online and offline training SLTrainer: supervised learning trainer pretrain, sft, dpo, reward model training OLTrainer: online learning trainer rlhf-ppo """ from abc import ABC, abstractmethod from contextlib import contextmanager from typing import Callable, List import torch.nn as nn import tqdm from coati.experience_buffer import NaiveExperienceBuffer from coati.experience_maker import Experience from torch.optim import Optimizer from colossalai.booster import Booster, Plugin from .utils import is_rank_0 class SLTrainer(ABC): """ Base class for supervised learning trainers. Args: strategy (Strategy):the strategy to use for training max_epochs (int, defaults to 1): the number of epochs of training process model (nn.Module): the model to train optim (Optimizer): the optimizer to use for training """ def __init__( self, booster: Booster, max_epochs: int, model: nn.Module, optimizer: Optimizer, plugin: Plugin, start_epoch: int = 0, ) -> None: super().__init__() self.booster = booster self.max_epochs = max_epochs self.model = model self.optimizer = optimizer self.plugin = plugin self.start_epoch = start_epoch @abstractmethod def _train(self, epoch): raise NotImplementedError() @abstractmethod def _eval(self, epoch): raise NotImplementedError() @abstractmethod def _before_fit(self): raise NotImplementedError() def fit(self, *args, **kwargs): self._before_fit(*args, **kwargs) for epoch in tqdm.trange(self.start_epoch, self.max_epochs, desc="Epochs", disable=not is_rank_0()): self._train(epoch) self._eval(epoch) class OLTrainer(ABC): """ Base class for online learning trainers, e.g. PPO. Args: strategy (Strategy):the strategy to use for training data_buffer (NaiveExperienceBuffer): the buffer to collect experiences sample_buffer (bool, defaults to False): whether to sample from buffer dataloader_pin_memory (bool, defaults to True): whether to pin memory for data loader callbacks (List[Callback], defaults to []): the callbacks to call during training process """ def __init__( self, actor_booster: Booster, critic_booster: Booster, data_buffer: NaiveExperienceBuffer, sample_buffer: bool, dataloader_pin_memory: bool, callbacks: List[Callable] = [], ) -> None: super().__init__() self.actor_booster = actor_booster self.critic_booster = critic_booster self.data_buffer = data_buffer self.sample_buffer = sample_buffer self.dataloader_pin_memory = dataloader_pin_memory self.callbacks = callbacks self.num_train_step = 0 @contextmanager def _fit_ctx(self) -> None: for callback in self.callbacks: callback.on_fit_start() try: yield finally: for callback in self.callbacks: callback.on_fit_end() @contextmanager def _episode_ctx(self, episode: int) -> None: for callback in self.callbacks: callback.on_episode_start(episode) try: yield finally: for callback in self.callbacks: callback.on_episode_end(episode) def _on_make_experience_start(self) -> None: for callback in self.callbacks: callback.on_make_experience_start() def _on_make_experience_end(self, experience: Experience) -> None: for callback in self.callbacks: callback.on_make_experience_end(experience) def _on_learn_epoch_start(self, epoch: int) -> None: for callback in self.callbacks: callback.on_learn_epoch_start(epoch) def _on_learn_epoch_end(self, epoch: int) -> None: for callback in self.callbacks: callback.on_learn_epoch_end(epoch) def _on_learn_batch_start(self) -> None: for callback in self.callbacks: callback.on_learn_batch_start() def _on_learn_batch_end(self, experience: Experience) -> None: for callback in self.callbacks: callback.on_learn_batch_end(experience) @abstractmethod def _make_experience(self, collect_step: int): """ Implement this method to make experience. """ raise NotImplementedError() @abstractmethod def _learn(self, update_step: int): """ Implement this method to learn from experience, either sample from buffer or transform buffer into dataloader. """ raise NotImplementedError() @abstractmethod def _setup_update_phrase_dataload(self): """ Implement this method to setup dataloader for update phase. """ raise NotImplementedError() @abstractmethod def _save_checkpoint(self, episode: int = 0): """ Implement this method to save checkpoint. """ raise NotImplementedError() def _collect_phase(self, collect_step: int): self._on_make_experience_start() experience = self._make_experience(collect_step) self._on_make_experience_end(experience) self.data_buffer.append(experience) def _update_phase(self, update_step: int): self._on_learn_epoch_start(update_step) self._learn(update_step) self._on_learn_epoch_end(update_step) def _before_fit(self, *args, **kwargs): raise NotImplementedError() def fit( self, num_episodes: int, num_collect_steps: int, num_update_steps: int, *args, **kwargs, ): """ The main training loop of on-policy rl trainers. Args: num_episodes (int): the number of episodes to train num_collect_steps (int): the number of collect steps per episode num_update_steps (int): the number of update steps per episode """ self._before_fit(*args, **kwargs) with self._fit_ctx(): for episode in tqdm.trange(num_episodes, desc="Episodes", disable=not is_rank_0()): with self._episode_ctx(episode): for collect_step in tqdm.trange(num_collect_steps, desc="Collect steps", disable=not is_rank_0()): self._collect_phase(collect_step) if not self.sample_buffer: self._setup_update_phrase_dataload() for update_step in tqdm.trange(num_update_steps, desc="Update steps", disable=not is_rank_0()): self._update_phase(update_step) # NOTE: this is for on-policy algorithms self.data_buffer.clear() if self.num_train_step > 0 and (self.num_train_step + 1) % (self.save_interval) == 0: self._save_checkpoint(self.num_train_step + 1)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/dpo.py
applications/ColossalChat/coati/trainer/dpo.py
""" Dpo trainer """ import os from typing import Any, Optional import torch import torch.distributed as dist from coati.models.loss import DpoLoss from coati.models.utils import calc_masked_log_probs from coati.trainer.utils import all_reduce_mean from coati.utils import AccumulativeMeanMeter, save_checkpoint from torch.optim import Optimizer from torch.optim.lr_scheduler import _LRScheduler from torch.utils.data import DataLoader from tqdm import tqdm, trange from transformers import PreTrainedTokenizerBase from colossalai.booster import Booster, Plugin from colossalai.booster.plugin import HybridParallelPlugin from colossalai.cluster import DistCoordinator from colossalai.utils import get_current_device from .base import SLTrainer from .utils import is_rank_0, to_device class DPOTrainer(SLTrainer): """ Trainer for DPO algorithm. Args: actor (Actor): the actor model in ppo algorithm ref_model (Critic): the reference model in ppo algorithm booster (Strategy): the strategy to use for training actor_optim (Optimizer): the optimizer to use for actor model actor_lr_scheduler (_LRScheduler): the lr scheduler to use for actor model tokenizer (PreTrainedTokenizerBase): the tokenizer to use for encoding max_epochs (int, defaults to 1): the max number of epochs to train beta (float, defaults to 0.1): the beta parameter in dpo loss accumulation_steps (int): the number of steps to accumulate gradients start_epoch (int, defaults to 0): the start epoch, non-zero if resumed from a checkpoint save_interval (int): the interval to save model checkpoints, default to 0, which means no checkpoint will be saved during trainning save_dir (str): the directory to save checkpoints coordinator (DistCoordinator): the coordinator to use for distributed logging """ def __init__( self, actor: Any, ref_model: Any, booster: Booster, actor_optim: Optimizer, plugin: Plugin, actor_lr_scheduler: _LRScheduler, tokenizer: PreTrainedTokenizerBase, max_epochs: int = 1, beta: float = 0.1, gamma: float = 0.0, length_normalization: bool = False, apply_loss_mask: bool = True, accumulation_steps: int = 1, start_epoch: int = 0, save_interval: int = 0, save_dir: str = None, coordinator: DistCoordinator = None, ) -> None: super().__init__( booster, max_epochs=max_epochs, model=actor, optimizer=actor_optim, plugin=plugin, start_epoch=start_epoch ) self.ref_model = ref_model self.actor_scheduler = actor_lr_scheduler self.tokenizer = tokenizer self.actor_loss_fn = DpoLoss(beta, gamma) self.apply_loss_mask = apply_loss_mask self.save_interval = save_interval self.coordinator = coordinator self.save_dir = save_dir self.num_train_step = 0 self.accumulation_steps = accumulation_steps self.device = get_current_device() self.accumulative_meter = AccumulativeMeanMeter() self.length_normalization = length_normalization def _before_fit( self, train_preference_dataloader: DataLoader = None, eval_preference_dataloader: DataLoader = None, log_dir: Optional[str] = None, use_wandb: bool = False, ): """ Args: prompt_dataloader (DataLoader): the dataloader to use for prompt data pretrain_dataloader (DataLoader): the dataloader to use for pretrain data """ self.train_dataloader = train_preference_dataloader self.eval_dataloader = eval_preference_dataloader self.writer = None init_criterion = ( dist.get_rank() == dist.get_world_size() - 1 if isinstance(self.plugin, HybridParallelPlugin) and self.plugin.pp_size > 1 else is_rank_0() ) if use_wandb and init_criterion: assert log_dir is not None, "log_dir must be provided when use_wandb is True" import wandb self.wandb_run = wandb.init(project="Coati-dpo", sync_tensorboard=True) if log_dir is not None and init_criterion: import os import time from torch.utils.tensorboard import SummaryWriter log_dir = os.path.join(log_dir, "DPO") log_dir = os.path.join(log_dir, time.strftime("%Y-%m-%d_%H:%M:%S", time.localtime())) self.writer = SummaryWriter(log_dir=log_dir) def _train(self, epoch: int): """ Args: epoch int: the number of current epoch """ self.model.train() if isinstance(self.plugin, HybridParallelPlugin) and self.plugin.pp_size > 1: step_bar = tqdm( range(len(self.train_dataloader)), desc="Step", disable=not (dist.get_rank() == dist.get_world_size() - 1), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) batch_size = chosen_input_ids.size()[0] # Calculate logits from reference model. if self.ref_model is not None: self.ref_model.eval() with torch.no_grad(): ref_all_logits = self.ref_model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] ref_chosen_logits = ref_all_logits[:batch_size] ref_reject_logits = ref_all_logits[batch_size:] logprob_ref_chosen = calc_masked_log_probs( ref_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_ref_reject = calc_masked_log_probs( ref_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) else: logprob_ref_chosen = None logprob_ref_reject = None # Merge chosen and reject inputs_ids = torch.stack([item for tup in zip(chosen_input_ids, reject_input_ids) for item in tup]) attention_mask = torch.stack( [item for tup in zip(chosen_attention_mask, reject_attention_mask) for item in tup] ) loss_mask = torch.stack([item for tup in zip(chosen_loss_mask, reject_loss_mask) for item in tup]) logprob_ref = torch.stack([item for tup in zip(logprob_ref_chosen, logprob_ref_reject) for item in tup]) data_iter = iter( [ { "input_ids": inputs_ids, "attention_mask": attention_mask, "loss_mask": loss_mask, "logprob_ref": logprob_ref, } ] ) rewards = [] def _criterion(outputs, inputs): loss, chosen_rewards, rejected_rewards = self.actor_loss_fn( calc_masked_log_probs( outputs["logits"][0::2], inputs["input_ids"][0::2], inputs["loss_mask"][0::2][:, 1:], self.length_normalization, ), calc_masked_log_probs( outputs["logits"][1::2], inputs["input_ids"][1::2], inputs["loss_mask"][1::2][:, 1:], self.length_normalization, ), inputs["logprob_ref"][0::2] if inputs["logprob_ref"] is not None else None, inputs["logprob_ref"][1::2] if inputs["logprob_ref"] is not None else None, inputs["loss_mask"][0::2][:, 1:], inputs["loss_mask"][1::2][:, 1:], ) rewards.append(chosen_rewards) rewards.append(rejected_rewards) return loss outputs = self.booster.execute_pipeline( data_iter, self.model, criterion=_criterion, optimizer=self.optimizer, return_loss=True, ) loss = outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): chosen_rewards, rejected_rewards = rewards[0], rewards[1] global_loss = all_reduce_mean(loss, self.plugin) if dist.get_rank() == dist.get_world_size() - 1: step_bar.set_postfix( { "train/loss": global_loss.item(), "train/lr": self.actor_scheduler.get_last_lr()[0], "train/chosen_rewards": chosen_rewards.to(torch.float16).mean().item(), "train/rejected_rewards": rejected_rewards.to(torch.float16).mean().item(), } ) step_bar.update() self.accumulative_meter.add("loss", global_loss.item()) self.accumulative_meter.add("chosen_rewards", chosen_rewards.to(torch.float16).mean().item()) self.accumulative_meter.add( "rejected_rewards", rejected_rewards.to(torch.float16).mean().item() ) if self.writer is not None: self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), i) self.writer.add_scalar( "train/chosen_rewards", self.accumulative_meter.get("chosen_rewards"), i ) self.writer.add_scalar( "train/rejected_rewards", self.accumulative_meter.get("rejected_rewards"), i, ) self.writer.add_scalar( "train/margin", self.accumulative_meter.get("chosen_rewards") - self.accumulative_meter.get("rejected_rewards"), i, ) self.optimizer.step() self.optimizer.zero_grad() self.actor_scheduler.step() else: self.accumulative_meter.reset() step_bar = trange( len(self.train_dataloader) // self.accumulation_steps, desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) for i, batch in enumerate(self.train_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) if not self.apply_loss_mask: chosen_loss_mask = chosen_loss_mask.fill_(1.0) reject_loss_mask = reject_loss_mask.fill_(1.0) batch_size = chosen_input_ids.size()[0] actor_all_logits = self.model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] actor_chosen_logits = actor_all_logits[:batch_size] actor_reject_logits = actor_all_logits[batch_size:] logprob_actor_chosen = calc_masked_log_probs( actor_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_actor_reject = calc_masked_log_probs( actor_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) if self.ref_model is not None: self.ref_model.eval() with torch.no_grad(): ref_all_logits = self.ref_model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] ref_chosen_logits = ref_all_logits[:batch_size] ref_reject_logits = ref_all_logits[batch_size:] logprob_ref_chosen = calc_masked_log_probs( ref_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_ref_reject = calc_masked_log_probs( ref_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) else: logprob_ref_chosen = None logprob_ref_reject = None loss, chosen_rewards, rejected_rewards = self.actor_loss_fn( logprob_actor_chosen, logprob_actor_reject, logprob_ref_chosen if logprob_ref_chosen is not None else None, logprob_ref_reject if logprob_ref_reject is not None else None, chosen_loss_mask[:, 1:], reject_loss_mask[:, 1:], ) reward_accuracies = (chosen_rewards > rejected_rewards).float().mean() self.booster.backward(loss=loss, optimizer=self.optimizer) # sync loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_rewards) rejected_rewards_mean = all_reduce_mean(tensor=rejected_rewards) reward_accuracies_mean = all_reduce_mean(tensor=reward_accuracies) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) self.accumulative_meter.add("accuracy", reward_accuracies_mean.to(torch.float16).item()) if (self.num_train_step + 1) % self.accumulation_steps == 0: self.optimizer.step() self.optimizer.zero_grad() self.actor_scheduler.step() step_bar.set_postfix( { "train/loss": self.accumulative_meter.get("loss"), "train/chosen_rewards": self.accumulative_meter.get("chosen_rewards"), "train/rejected_rewards": self.accumulative_meter.get("rejected_rewards"), "train/accuracy": self.accumulative_meter.get("accuracy"), } ) step_bar.update() if self.writer and is_rank_0(): global_step = (self.num_train_step + 1) / self.accumulation_steps self.writer.add_scalar("train/loss", self.accumulative_meter.get("loss"), global_step) self.writer.add_scalar("train/lr", self.optimizer.param_groups[0]["lr"], global_step) self.writer.add_scalar( "train/chosen_rewards", self.accumulative_meter.get("chosen_rewards"), global_step ) self.writer.add_scalar( "train/rejected_rewards", self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/margin", self.accumulative_meter.get("chosen_rewards") - self.accumulative_meter.get("rejected_rewards"), global_step, ) self.writer.add_scalar( "train/accuracy", self.accumulative_meter.get("accuracy"), global_step, ) self.accumulative_meter.reset() self.num_train_step += 1 if self.save_dir is not None and self.num_train_step > 0 and self.num_train_step % self.save_interval == 0: # save checkpoint self.coordinator.print_on_master("\nStart saving model checkpoint with running states") save_checkpoint( save_dir=self.save_dir, booster=self.booster, model=self.model, optimizer=self.optimizer, lr_scheduler=self.actor_scheduler, epoch=epoch, step=self.num_train_step, batch_size=batch_size, coordinator=self.coordinator, ) self.coordinator.print_on_master( f"Saved checkpoint at epoch {epoch} step {self.save_interval} at folder {self.save_dir}" ) step_bar.close() def _eval(self, epoch: int): """ Args: epoch int: the number of current epoch """ if self.eval_dataloader is None: self.coordinator.print_on_master("No eval dataloader is provided, skip evaluation") return self.model.eval() self.ref_model.eval() self.accumulative_meter.reset() self.coordinator.print_on_master("\nStart evaluation...") if isinstance(self.plugin, HybridParallelPlugin) and self.plugin.pp_size > 1: step_bar = tqdm( range(len(self.eval_dataloader)), desc="Step", disable=not (dist.get_rank() == dist.get_world_size() - 1), ) with torch.no_grad(): for _, batch in enumerate(self.eval_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) batch_size = chosen_input_ids.size()[0] # Calculate logits from reference model. if self.ref_model is not None: self.ref_model.eval() with torch.no_grad(): ref_all_logits = self.ref_model( input_ids=torch.cat([chosen_input_ids, reject_input_ids]), attention_mask=torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] ref_chosen_logits = ref_all_logits[:batch_size] ref_reject_logits = ref_all_logits[batch_size:] logprob_ref_chosen = calc_masked_log_probs( ref_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_ref_reject = calc_masked_log_probs( ref_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) else: logprob_ref_chosen = None logprob_ref_reject = None # Merge chosen and reject inputs_ids = torch.stack([item for tup in zip(chosen_input_ids, reject_input_ids) for item in tup]) attention_mask = torch.stack( [item for tup in zip(chosen_attention_mask, reject_attention_mask) for item in tup] ) loss_mask = torch.stack([item for tup in zip(chosen_loss_mask, reject_loss_mask) for item in tup]) logprob_ref = torch.stack( [item for tup in zip(logprob_ref_chosen, logprob_ref_reject) for item in tup] ) data_iter = iter( [ { "input_ids": inputs_ids, "attention_mask": attention_mask, "loss_mask": loss_mask, "logprob_ref": logprob_ref, } ] ) rewards = [] def _criterion(outputs, inputs): loss, chosen_rewards, rejected_rewards = self.actor_loss_fn( calc_masked_log_probs( outputs["logits"][0::2], inputs["input_ids"][0::2], inputs["loss_mask"][0::2][:, 1:], self.length_normalization, ), calc_masked_log_probs( outputs["logits"][1::2], inputs["input_ids"][1::2], inputs["loss_mask"][1::2][:, 1:], self.length_normalization, ), inputs["logprob_ref"][0::2] if inputs["logprob_ref"] is not None else None, inputs["logprob_ref"][1::2] if inputs["logprob_ref"] is not None else None, inputs["loss_mask"][0::2][:, 1:], inputs["loss_mask"][1::2][:, 1:], ) rewards.append(chosen_rewards) rewards.append(rejected_rewards) return loss outputs = self.booster.execute_pipeline( data_iter, self.model, criterion=_criterion, optimizer=self.optimizer, return_loss=True, ) loss = outputs["loss"] if self.booster.plugin.stage_manager.is_last_stage(): chosen_rewards, rejected_rewards = rewards[0], rewards[1] global_loss = all_reduce_mean(loss, self.plugin) chosen_rewards_mean = all_reduce_mean(chosen_rewards, self.plugin) rejected_rewards_mean = all_reduce_mean(rejected_rewards, self.plugin) if dist.get_rank() == dist.get_world_size() - 1: step_bar.set_postfix( { "eval/loss": global_loss.item(), "eval/lr": self.actor_scheduler.get_last_lr()[0], "eval/chosen_rewards": chosen_rewards.to(torch.float16).mean().item(), "eval/rejected_rewards": rejected_rewards.to(torch.float16).mean().item(), } ) self.accumulative_meter.add( "chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item() ) self.accumulative_meter.add( "rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item() ) self.accumulative_meter.add("loss", global_loss.to(torch.float16).item()) step_bar.update() if self.booster.plugin.stage_manager.is_last_stage(): msg = "\nEvaluation Result:\n" for tag in ["loss", "chosen_rewards", "rejected_rewards"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" if dist.get_rank() == dist.get_world_size() - 1: print(msg) else: step_bar = trange( len(self.eval_dataloader), desc=f"Epoch {epoch + 1}/{self.max_epochs}", disable=not is_rank_0(), ) with torch.no_grad(): for i, batch in enumerate(self.eval_dataloader): batch = to_device(batch, self.device) ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( batch["chosen_input_ids"], batch["chosen_attention_mask"], batch["chosen_loss_mask"], batch["reject_input_ids"], batch["reject_attention_mask"], batch["reject_loss_mask"], ) if not self.apply_loss_mask: chosen_loss_mask = chosen_loss_mask.fill_(1.0) reject_loss_mask = reject_loss_mask.fill_(1.0) batch_size = chosen_input_ids.size()[0] actor_all_logits = self.model( torch.cat([chosen_input_ids, reject_input_ids]), torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] actor_chosen_logits = actor_all_logits[:batch_size] actor_reject_logits = actor_all_logits[batch_size:] logprob_actor_chosen = calc_masked_log_probs( actor_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_actor_reject = calc_masked_log_probs( actor_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) ref_all_logits = self.ref_model( torch.cat([chosen_input_ids, reject_input_ids]), torch.cat([chosen_attention_mask, reject_attention_mask]), )["logits"] ref_chosen_logits = ref_all_logits[:batch_size] ref_reject_logits = ref_all_logits[batch_size:] logprob_ref_chosen = calc_masked_log_probs( ref_chosen_logits, chosen_input_ids, chosen_loss_mask[:, 1:], self.length_normalization ) logprob_ref_reject = calc_masked_log_probs( ref_reject_logits, reject_input_ids, reject_loss_mask[:, 1:], self.length_normalization ) losses, chosen_rewards, rejected_rewards = self.actor_loss_fn( logprob_actor_chosen, logprob_actor_reject, logprob_ref_chosen if logprob_ref_chosen is not None else None, logprob_ref_reject if logprob_ref_reject is not None else None, chosen_loss_mask[:, 1:], reject_loss_mask[:, 1:], ) reward_accuracies = (chosen_rewards > rejected_rewards).float().mean() loss = losses.mean() loss_mean = all_reduce_mean(tensor=loss) chosen_rewards_mean = all_reduce_mean(tensor=chosen_rewards) rejected_rewards_mean = all_reduce_mean(tensor=rejected_rewards) reward_accuracies_mean = all_reduce_mean(tensor=reward_accuracies) self.accumulative_meter.add("chosen_rewards", chosen_rewards_mean.to(torch.float16).mean().item()) self.accumulative_meter.add( "rejected_rewards", rejected_rewards_mean.to(torch.float16).mean().item() ) self.accumulative_meter.add("loss", loss_mean.to(torch.float16).item()) self.accumulative_meter.add("accuracy", reward_accuracies_mean.to(torch.float16).item()) self.accumulative_meter.add( "margin", (chosen_rewards_mean - rejected_rewards_mean).to(torch.float16).mean().item() ) step_bar.set_postfix( { "eval/loss": self.accumulative_meter.get("loss"), "eval/chosen_rewards": self.accumulative_meter.get("chosen_rewards"), "eval/rejected_rewards": self.accumulative_meter.get("rejected_rewards"), "eval/accuracy": self.accumulative_meter.get("accuracy"), } ) step_bar.update() msg = "\nEvaluation Result:\n" for tag in ["loss", "chosen_rewards", "rejected_rewards", "accuracy", "margin"]: msg = msg + f"{tag}: {self.accumulative_meter.get(tag)}\n" self.coordinator.print_on_master(msg) if self.save_dir is not None: os.makedirs(self.save_dir, exist_ok=True) with open(os.path.join(self.save_dir, f"eval_result_epoch{epoch}.txt"), "w") as f: f.write(msg) step_bar.close()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/callbacks/__init__.py
applications/ColossalChat/coati/trainer/callbacks/__init__.py
from .base import Callback from .performance_evaluator import PerformanceEvaluator __all__ = ["Callback", "PerformanceEvaluator"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/callbacks/base.py
applications/ColossalChat/coati/trainer/callbacks/base.py
from abc import ABC from coati.experience_maker import Experience class Callback(ABC): """ Base callback class. It defines the interface for callbacks. """ def on_fit_start(self) -> None: pass def on_fit_end(self) -> None: pass def on_episode_start(self, episode: int) -> None: pass def on_episode_end(self, episode: int) -> None: pass def on_make_experience_start(self) -> None: pass def on_make_experience_end(self, experience: Experience) -> None: pass def on_learn_epoch_start(self, epoch: int) -> None: pass def on_learn_epoch_end(self, epoch: int) -> None: pass def on_learn_batch_start(self) -> None: pass def on_learn_batch_end(self, experience: Experience) -> None: pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/trainer/callbacks/performance_evaluator.py
applications/ColossalChat/coati/trainer/callbacks/performance_evaluator.py
from time import time from typing import Optional import torch import torch.distributed as dist from coati.experience_maker import Experience from .base import Callback def get_world_size() -> int: if dist.is_initialized(): return dist.get_world_size() return 1 def save_eval_result_rank_0(s: str, save_path: str, **kwargs) -> None: if not dist.is_initialized() or dist.get_rank() == 0: with open(save_path, "a+") as f: train_config = "; ".join([str(kwargs[key]) for key in kwargs]) f.write(train_config + "\n" + s + "\n") def divide(x: float, y: float) -> float: if y == 0: return float("inf") elif y == float("inf"): return float("nan") return x / y @torch.no_grad() def all_reduce_mean(x: float, world_size: int) -> float: if world_size == 1: return x tensor = torch.tensor([x], device=torch.cuda.current_device()) dist.all_reduce(tensor) tensor = tensor / world_size return tensor.item() class Timer: def __init__(self) -> None: self.start_time: Optional[float] = None self.duration: float = 0.0 def start(self) -> None: self.start_time = time() def end(self) -> None: assert self.start_time is not None self.duration += time() - self.start_time self.start_time = None def reset(self) -> None: self.duration = 0.0 class PerformanceEvaluator(Callback): """ Callback for valuate the performance of the model. Args: actor_num_params: The number of parameters of the actor model. critic_num_params: The number of parameters of the critic model. initial_model_num_params: The number of parameters of the initial model. reward_model_num_params: The number of parameters of the reward model. enable_grad_checkpoint: Whether to enable gradient checkpointing. ignore_episodes: The number of episodes to ignore when calculating the performance. """ def __init__( self, actor_num_params: int, critic_num_params: int, initial_model_num_params: int, reward_model_num_params: int, enable_grad_checkpoint: bool = False, ignore_episodes: int = 0, train_config: Optional[dict] = None, save_path: Optional[str] = None, ) -> None: super().__init__() self.world_size = get_world_size() self.actor_num_params = actor_num_params self.critic_num_params = critic_num_params self.initial_model_num_params = initial_model_num_params self.reward_model_num_params = reward_model_num_params self.enable_grad_checkpoint = enable_grad_checkpoint self.ignore_episodes = ignore_episodes self.disable: bool = False self.overall_timer = Timer() self.make_experience_timer = Timer() self.learn_timer = Timer() self.make_experience_num_samples: int = 0 self.make_experience_flop: int = 0 self.learn_num_samples: int = 0 self.learn_flop: int = 0 self.train_config = train_config self.save_path = save_path def on_episode_start(self, episode: int) -> None: self.disable = self.ignore_episodes > 0 and episode < self.ignore_episodes if self.disable: return self.overall_timer.start() def on_episode_end(self, episode: int) -> None: if self.disable: return self.overall_timer.end() def on_make_experience_start(self) -> None: if self.disable: return self.make_experience_timer.start() def on_make_experience_end(self, experience: Experience) -> None: if self.disable: return self.make_experience_timer.end() batch_size, seq_len = experience.sequences.shape self.make_experience_num_samples += batch_size # actor generate num_actions = experience.action_mask.size(1) input_len = seq_len - num_actions total_seq_len = (input_len + seq_len - 1) * num_actions / 2 self.make_experience_flop += self.actor_num_params * batch_size * total_seq_len * 2 # actor forward self.make_experience_flop += self.actor_num_params * batch_size * seq_len * 2 # critic forward self.make_experience_flop += self.critic_num_params * batch_size * seq_len * 2 # initial model forward self.make_experience_flop += self.initial_model_num_params * batch_size * seq_len * 2 # reward model forward self.make_experience_flop += self.reward_model_num_params * batch_size * seq_len * 2 def on_learn_batch_start(self) -> None: if self.disable: return self.learn_timer.start() def on_learn_batch_end(self, experience: Experience) -> None: if self.disable: return self.learn_timer.end() batch_size, seq_len = experience.sequences.shape self.learn_num_samples += batch_size # actor forward-backward, 3 means forward(1) + backward(2) self.learn_flop += self.actor_num_params * batch_size * seq_len * 2 * (3 + int(self.enable_grad_checkpoint)) # critic forward-backward self.learn_flop += self.critic_num_params * batch_size * seq_len * 2 * (3 + int(self.enable_grad_checkpoint)) def on_fit_end(self) -> None: avg_make_experience_duration = all_reduce_mean(self.make_experience_timer.duration, self.world_size) avg_learn_duration = all_reduce_mean(self.learn_timer.duration, self.world_size) avg_overall_duration = all_reduce_mean(self.overall_timer.duration, self.world_size) avg_make_experience_throughput = ( self.make_experience_num_samples * self.world_size / (avg_make_experience_duration + 1e-12) ) avg_make_experience_tflops = self.make_experience_flop / 1e12 / (avg_make_experience_duration + 1e-12) avg_learn_throughput = self.learn_num_samples * self.world_size / (avg_learn_duration + 1e-12) avg_learn_tflops = self.learn_flop / 1e12 / (avg_learn_duration + 1e-12) num_effective_samples = min(self.learn_num_samples, self.make_experience_num_samples) * self.world_size avg_overall_throughput = num_effective_samples / (avg_overall_duration + 1e-12) overall_time_per_sample = divide(1, avg_overall_throughput) make_experience_time_per_sample = divide(avg_make_experience_duration, num_effective_samples) learn_time_per_sample = divide(avg_learn_duration, num_effective_samples) save_eval_result_rank_0( f"Performance summary:\n" + f"Generate {self.make_experience_num_samples * self.world_size} samples, throughput: {avg_make_experience_throughput:.2f} samples/s, TFLOPS per GPU: {avg_make_experience_tflops:.2f}\n" + f"Train {self.learn_num_samples * self.world_size} samples, throughput: {avg_learn_throughput:.2f} samples/s, TFLOPS per GPU: {avg_learn_tflops:.2f}\n" + f"Overall throughput: {avg_overall_throughput:.2f} samples/s\n" + f"Overall time per sample: {overall_time_per_sample:.2f} s\n" + f"Make experience time per sample: {make_experience_time_per_sample:.2f} s, {make_experience_time_per_sample/overall_time_per_sample*100:.2f}%\n" + f"Learn time per sample: {learn_time_per_sample:.2f} s, {learn_time_per_sample/overall_time_per_sample*100:.2f}%", self.save_path, **self.train_config, )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/accumulative_meter.py
applications/ColossalChat/coati/utils/accumulative_meter.py
""" A class that can be used to calculate the mean of a variable """ class AccumulativeMeanVariable: """ A class that calculates the accumulative mean of a variable. """ def __init__(self): self._sum = 0 self._count = 0 def add(self, value, count_update=1): """ Adds a value to the sum and updates the count. Args: value (float): The value to be added. count_update (int, optional): The amount to update the count by. Defaults to 1. """ self._sum += value self._count += count_update def get(self): """ Calculates and returns the accumulative mean. Returns: float: The accumulative mean. """ return self._sum / self._count if self._count > 0 else 0 def reset(self): """ Resets the sum and count to zero. """ self._sum = 0 self._count = 0 class AccumulativeMeanMeter: """ A class for calculating and storing the accumulative mean of variables. Attributes: variable_dict (dict): A dictionary to store the accumulative mean variables. Methods: add(name, value, count_update=1): Adds a value to the specified variable. get(name): Retrieves the accumulative mean value of the specified variable. reset(): Resets all the accumulative mean variables to their initial state. """ def __init__(self): self.variable_dict = {} def add(self, name, value, count_update=1): if name not in self.variable_dict: self.variable_dict[name] = AccumulativeMeanVariable() self.variable_dict[name].add(value, count_update=count_update) def get(self, name): return self.variable_dict[name].get() def reset(self): for name in self.variable_dict: self.variable_dict[name].reset()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/ckpt_io.py
applications/ColossalChat/coati/utils/ckpt_io.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Helper functions for IO save load checkpoints """ import json import os from typing import Any, Dict, Tuple, Union import torch from torch.optim.lr_scheduler import _LRScheduler from torch.optim.optimizer import Optimizer from colossalai.booster import Booster from colossalai.cluster import DistCoordinator def load_json(file_path: Union[str, os.PathLike]) -> Dict[str, Any]: """ Load file in JSON format """ with open(file=file_path, mode="r", encoding="utf-8") as fp: return json.load(fp) def save_json(data: Dict[str, Any], file_path: Union[str, os.PathLike]) -> None: """ Save as JSON format """ with open(file=file_path, mode="w", encoding="utf-8") as fp: json.dump(data, fp=fp, ensure_ascii=False, indent=4) def save_checkpoint( save_dir: Union[str, os.PathLike], booster: Booster, model: torch.nn.Module, optimizer: Optimizer, lr_scheduler: _LRScheduler, epoch: int, step: int, batch_size: int, coordinator: DistCoordinator, ) -> None: """ Save model checkpoint, optimizer, LR scheduler and intermedidate running states. """ save_dir = os.path.join(save_dir, f"epoch-{epoch}_step-{step}") os.makedirs(os.path.join(save_dir, "modeling"), exist_ok=True) booster.save_model(model, os.path.join(save_dir, "modeling"), shard=True) """ Temporary disable the following as save_optimizer causes all processes to hang in a multi-gpu environment, working on fixing this bug """ booster.save_optimizer(optimizer, os.path.join(save_dir, "optimizer"), shard=True) booster.save_lr_scheduler(lr_scheduler, os.path.join(save_dir, "lr_scheduler")) running_states = { "epoch": epoch, "step": step, "sample_start_index": step * batch_size, } if coordinator.is_master(): save_json(running_states, os.path.join(save_dir, "running_states.json")) def load_checkpoint( load_dir: Union[str, os.PathLike], booster: Booster, model: torch.nn.Module, optimizer: Optimizer, lr_scheduler: _LRScheduler, ) -> Tuple[int, int, int]: """ Load model checkpoint, optimizer, LR scheduler and intermedidate running states. """ # Update booster params states. if model is not None: booster.load_model(model=model, checkpoint=os.path.join(load_dir, "modeling")) if optimizer is not None: booster.load_optimizer(optimizer=optimizer, checkpoint=os.path.join(load_dir, "optimizer")) if lr_scheduler is not None: booster.load_lr_scheduler(lr_scheduler=lr_scheduler, checkpoint=os.path.join(load_dir, "lr_scheduler")) running_states = load_json(file_path=os.path.join(load_dir, "running_states.json")) return ( running_states["epoch"], running_states["step"], running_states["sample_start_index"], )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/__init__.py
applications/ColossalChat/coati/utils/__init__.py
from .accumulative_meter import AccumulativeMeanMeter from .ckpt_io import load_checkpoint, save_checkpoint __all__ = ["load_checkpoint", "save_checkpoint", "AccumulativeMeanMeter"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/reward_score/gsm8k.py
applications/ColossalChat/coati/utils/reward_score/gsm8k.py
import torch from .utils import extract_solution, validate_response_structure def gsm8k_reward_fn(input_ids, attention_mask, **kwargs): # apply varifiable reward # reward 10 points if the final answer is correct, reward 1 point if format is correct gt_answer = kwargs["gt_answer"] tokenizer = kwargs["tokenizer"] s, e = kwargs["response_start"], kwargs["response_end"] reward = torch.tensor(0.0).to(input_ids.device) if gt_answer is None: return reward decoded_final_answer = tokenizer.decode(input_ids[s:e], skip_special_tokens=True) final_answer, processed_str = extract_solution(decoded_final_answer) is_valid = True try: int(final_answer.strip()) except Exception: is_valid = False format_valid = validate_response_structure(processed_str, kwargs["tags"]) if not is_valid or not format_valid: return reward else: reward += 1.0 if gt_answer.strip().replace(" ", "").lower() == final_answer.strip().replace(" ", "").lower(): reward = reward + 9.0 return reward
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/reward_score/utils.py
applications/ColossalChat/coati/utils/reward_score/utils.py
# Copyright Unakar # Modified from https://github.com/Unakar/Logic-RL/blob/086373176ac198c97277ff50f4b6e7e1bfe669d3/verl/utils/reward_score/kk.py#L99 # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from typing import Dict, Optional, Tuple def validate_response_structure(processed_str: str, tags: Dict = None) -> bool: """Performs comprehensive validation of response structure. Args: processed_str: Processed response string from the model Returns: Boolean indicating whether all formatting requirements are met """ validation_passed = True # Check required tags if tags is None: tags = { "think_start": {"text": "<think>", "num_occur": 1}, "think_end": {"text": "</think>", "num_occur": 1}, "answer_start": {"text": "<answer>", "num_occur": 1}, "answer_end": {"text": "</answer>", "num_occur": 1}, } positions = {} for tag_name, tag_info in tags.items(): tag_str = tag_info["text"] expected_count = tag_info["num_occur"] count = processed_str.count(tag_str) positions[tag_name] = pos = processed_str.find(tag_str) if count != expected_count: validation_passed = False # Verify tag order if ( positions["think_start"] > positions["think_end"] or positions["think_end"] > positions["answer_start"] or positions["answer_start"] > positions["answer_end"] ): validation_passed = False if len(processed_str) - positions["answer_end"] != len(tags["answer_end"]["text"]): validation_passed = False return validation_passed def extract_solution(solution_str: str) -> Tuple[Optional[str], str]: """Extracts the final answer from the model's response string. Args: solution_str: Raw response string from the language model Returns: Tuple containing (extracted_answer, processed_string) """ # Extract final answer using XML-style tags answer_pattern = r"<answer>(.*?)</answer>" matches = list(re.finditer(answer_pattern, solution_str, re.DOTALL)) if not matches: return None, solution_str final_answer = matches[-1].group(1).strip() return final_answer, solution_str
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/reward_score/__init__.py
applications/ColossalChat/coati/utils/reward_score/__init__.py
from .competition import math_competition_reward_fn from .gsm8k import gsm8k_reward_fn __all__ = ["gsm8k_reward_fn", "math_competition_reward_fn"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/utils/reward_score/competition.py
applications/ColossalChat/coati/utils/reward_score/competition.py
import torch from .utils import extract_solution, validate_response_structure def math_competition_reward_fn(input_ids, attention_mask, **kwargs): # apply varifiable reward # reward 10 points if the final answer is correct, reward 1 point if format is correct gt_answer = kwargs["gt_answer"] tokenizer = kwargs["tokenizer"] s, e = kwargs["response_start"], kwargs["response_end"] reward = torch.tensor(0.0).to(input_ids.device) if gt_answer is None: return reward decoded_final_answer = tokenizer.decode(input_ids[s : e + 1], skip_special_tokens=True) final_answer, processed_str = extract_solution(decoded_final_answer) format_valid = validate_response_structure(processed_str, kwargs["tags"]) if not format_valid: return reward else: reward += 1.0 if gt_answer.strip().replace(" ", "").lower() == final_answer.strip().replace(" ", "").lower(): reward = reward + 9.0 return reward
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/dataset/loader.py
applications/ColossalChat/coati/dataset/loader.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Dataloader for sft, dpo, ppo """ import os from dataclasses import dataclass from typing import Dict, Iterator, List, Optional, Sequence, Union import jsonlines import torch import torch.nn.functional as F from coati.dataset.utils import chuncate_sequence, pad_to_max_len from datasets import Dataset as HFDataset from datasets import dataset_dict, load_from_disk from torch.utils.data import ConcatDataset, Dataset, DistributedSampler from transformers.tokenization_utils import PreTrainedTokenizer DatasetType = Union[Dataset, ConcatDataset, dataset_dict.Dataset] PathType = Union[str, os.PathLike] def load_tokenized_dataset( dataset_paths: Union[PathType, List[PathType]], mode: str = "train", **kwargs ) -> Optional[DatasetType]: """ Load pre-tokenized dataset. Each instance of dataset is a dictionary with `{'input_ids': List[int], 'labels': List[int], sequence: str}` format. """ if not dataset_paths: return None mode_map = kwargs.get("mode_map", {"train": "train", "dev": "validation", "test": "test"}) assert mode in tuple(mode_map), f"Unsupported mode {mode}, it must be in {tuple(mode_map)}" if isinstance(dataset_paths, (str, os.PathLike)): dataset_paths = [dataset_paths] datasets = [] # `List[datasets.dataset_dict.Dataset]` for ds_path in dataset_paths: ds_path = os.path.abspath(ds_path) assert os.path.exists(ds_path), f"Not existed file path {ds_path}" ds_dict = load_from_disk(dataset_path=ds_path, keep_in_memory=False) if isinstance(ds_dict, HFDataset): datasets.append(ds_dict) else: if mode_map[mode] in ds_dict: datasets.append(ds_dict[mode_map[mode]]) if len(datasets) == 0: return None if len(datasets) == 1: return datasets.pop() return ConcatDataset(datasets=datasets) @dataclass class DataCollatorForSupervisedDataset(object): """ Collate instances for supervised dataset. Each instance is a tokenized dictionary with fields `input_ids`(List[int]), `labels`(List[int]) and `sequence`(str). """ tokenizer: PreTrainedTokenizer max_length: int = 4096 ignore_index: int = -100 def __call__(self, instances: Sequence[Dict[str, List[int]]]) -> Dict[str, torch.Tensor]: """ Args: instances (`Sequence[Dict[str, List[int]]]`): Mini-batch samples, each sample is stored in an individual dictionary. Returns: (`Dict[str, torch.Tensor]`): Contains the following `torch.Tensor`: `input_ids`: `torch.Tensor` of shape (bsz, max_len); `attention_mask`: `torch.BoolTensor` of shape (bsz, max_len); `labels`: `torch.Tensor` of shape (bsz, max_len), which contains `IGNORE_INDEX`. """ assert isinstance(self.tokenizer.pad_token_id, int) and self.tokenizer.pad_token_id >= 0, ( f"`{self.tokenizer.__class__.__name__}.pad_token_id` must be a valid non-negative integer index value, " f"but now `{self.tokenizer.pad_token_id}`" ) # `List[torch.Tensor]` batch_input_ids = [ ( torch.LongTensor(instance["input_ids"][: self.max_length]) if len(instance["input_ids"]) > self.max_length else torch.LongTensor(instance["input_ids"]) ) for instance in instances ] batch_labels = [ ( torch.LongTensor(instance["labels"][: self.max_length]) if len(instance["labels"]) > self.max_length else torch.LongTensor(instance["labels"]) ) for instance in instances ] if self.tokenizer.padding_side == "right": input_ids = torch.nn.utils.rnn.pad_sequence( sequences=batch_input_ids, batch_first=True, padding_value=self.tokenizer.pad_token_id, ) # (bsz, max_len) labels = torch.nn.utils.rnn.pad_sequence( sequences=batch_labels, batch_first=True, padding_value=self.ignore_index, ) # (bsz, max_len) # pad to max to_pad = self.max_length - input_ids.size(1) input_ids = F.pad(input_ids, (0, to_pad), value=self.tokenizer.pad_token_id) labels = F.pad(labels, (0, to_pad), value=self.ignore_index) elif self.tokenizer.padding_side == "left": reversed_input_ids = [seq.flip(dims=(0,)) for seq in batch_input_ids] reversed_input_ids = torch.nn.utils.rnn.pad_sequence( sequences=reversed_input_ids, batch_first=True, padding_value=self.tokenizer.pad_token_id, ) # (bsz, max_len) input_ids = torch.flip(reversed_input_ids, dims=(1,)) # (bsz, max_len) reversed_labels = [seq.flip(dims=(0,)) for seq in batch_labels] reversed_labels = torch.nn.utils.rnn.pad_sequence( sequences=reversed_labels, batch_first=True, padding_value=self.ignore_index, ) # (bsz, max_len) labels = torch.flip(reversed_labels, dims=(1,)) # (bsz, max_len) else: raise RuntimeError( f"`{self.tokenizer.__class__.__name__}.padding_side` can only be `left` or `right`, " f"but now `{self.tokenizer.padding_side}`" ) attention_mask = input_ids.ne(self.tokenizer.pad_token_id) # `torch.BoolTensor`, (bsz, max_len) return dict(input_ids=input_ids, attention_mask=attention_mask, labels=labels) @dataclass class DataCollatorForPromptDataset(DataCollatorForSupervisedDataset): def __call__(self, instances: Sequence[Dict[str, List[int]]]) -> Dict[str, torch.Tensor]: """ Args: instances (`Sequence[Dict[str, List[int]]]`): Mini-batch samples, each sample is stored in an individual dictionary. Returns: (`Dict[str, torch.Tensor]`): Contains the following `torch.Tensor`: `input_ids`: `torch.Tensor` of shape (bsz, max_len); `attention_mask`: `torch.BoolTensor` of shape (bsz, max_len); """ gt_answer = [ins.get("gt_answer", None) for ins in instances] instances = [{"input_ids": ins["input_ids"], "labels": ins["input_ids"]} for ins in instances] ret = super().__call__(instances=instances) input_ids = F.pad( ret["input_ids"], (self.max_length - ret["input_ids"].size(1), 0), value=self.tokenizer.pad_token_id ) attention_mask = F.pad(ret["attention_mask"], (self.max_length - ret["attention_mask"].size(1), 0), value=False) return {"input_ids": input_ids, "attention_mask": attention_mask, "gt_answer": gt_answer} @dataclass class DataCollatorForPreferenceDataset(object): """ Collate instances for supervised dataset. Each instance is a tokenized dictionary with fields `input_ids`(List[int]), `labels`(List[int]) and `sequence`(str). """ tokenizer: PreTrainedTokenizer max_length: int = 4096 def __call__(self, instances: Sequence[Dict[str, List[int]]]) -> Dict[str, torch.Tensor]: """ Args: instances (`Sequence[Dict[str, List[int]]]`): Mini-batch samples, each sample is stored in an individual dictionary. Returns: (`Dict[str, torch.Tensor]`): Contains the following `torch.Tensor`: `input_ids`: `torch.Tensor` of shape (bsz, max_len); `attention_mask`: `torch.BoolTensor` of shape (bsz, max_len); `labels`: `torch.Tensor` of shape (bsz, max_len), which contains `IGNORE_INDEX`. """ assert isinstance(self.tokenizer.pad_token_id, int) and self.tokenizer.pad_token_id >= 0, ( f"`{self.tokenizer.__class__.__name__}.pad_token_id` must be a valid non-negative integer index value, " f"but now `{self.tokenizer.pad_token_id}`" ) ( chosen_input_ids, chosen_loss_mask, # [batch_size * seq_len] reject_input_ids, reject_loss_mask, ) = ( chuncate_sequence([ins["chosen_input_ids"] for ins in instances], self.max_length, torch.int64), chuncate_sequence([ins["chosen_loss_mask"] for ins in instances], self.max_length, torch.bool), chuncate_sequence([ins["rejected_input_ids"] for ins in instances], self.max_length, torch.int64), chuncate_sequence([ins["rejected_loss_mask"] for ins in instances], self.max_length, torch.bool), ) padding_side = self.tokenizer.padding_side chosen_attention_mask = [torch.ones_like(seq).bool() for seq in chosen_input_ids] reject_attention_mask = [torch.ones_like(seq).bool() for seq in reject_input_ids] ( chosen_input_ids, chosen_attention_mask, chosen_loss_mask, reject_input_ids, reject_attention_mask, reject_loss_mask, ) = ( pad_to_max_len(chosen_input_ids, self.max_length, self.tokenizer.pad_token_id, padding_side=padding_side), pad_to_max_len(chosen_attention_mask, self.max_length, False, padding_side=padding_side), pad_to_max_len(chosen_loss_mask, self.max_length, False, padding_side=padding_side), pad_to_max_len(reject_input_ids, self.max_length, self.tokenizer.pad_token_id, padding_side=padding_side), pad_to_max_len(reject_attention_mask, self.max_length, False, padding_side=padding_side), pad_to_max_len(reject_loss_mask, self.max_length, False, padding_side=padding_side), ) return dict( chosen_input_ids=chosen_input_ids, chosen_attention_mask=chosen_attention_mask, chosen_loss_mask=chosen_loss_mask, reject_input_ids=reject_input_ids, reject_attention_mask=reject_attention_mask, reject_loss_mask=reject_loss_mask, ) @dataclass class DataCollatorForKTODataset(object): """ Collate instances for kto dataset. Each input instance is a tokenized dictionary with fields `prompt`(List[int]), `completion`(List[int]) and `label`(bool). Each output instance is a tokenized dictionary with fields `kl_input_ids`(List[int]), `kl_attention_mask`(List[int]) and `kl_loss_mask`(List[int]). `input_ids`(List[int]), `attention_mask`(List[int]), `loss_mask`(List[int]) and `label`(bool). """ tokenizer: PreTrainedTokenizer max_length: int = 4096 ignore_index: int = -100 def __call__(self, instances: Sequence[Dict[str, List[int]]]) -> Dict[str, torch.Tensor]: """ Args: instances (`Sequence[Dict[str, List[int]]]`): Mini-batch samples, each sample is stored in an individual dictionary contains the following fields: `prompt`(List[int]), `completion`(List[int]) and `label`(bool, if the sample is desirable or not). Returns: (`Dict[str, torch.Tensor]`): Contains the following `torch.Tensor`: `input_ids`: `torch.Tensor` of shape (bsz, max_len); `attention_mask`: `torch.BoolTensor` of shape (bsz, max_len); `labels`: `torch.Tensor` of shape (bsz, max_len), which contains `IGNORE_INDEX`. """ assert isinstance(self.tokenizer.pad_token_id, int) and self.tokenizer.pad_token_id >= 0, ( f"`{self.tokenizer.__class__.__name__}.pad_token_id` must be a valid non-negative integer index value, " f"but now `{self.tokenizer.pad_token_id}`" ) # prepare the preference data prompt = [torch.LongTensor(instance["prompt"]) for instance in instances] prompt_zeros = [torch.zeros_like(t) for t in prompt] completion = [torch.LongTensor(instance["completion"]) for instance in instances] completion_ones = [torch.ones_like(t) for t in completion] label = [torch.tensor(instance["label"], dtype=torch.bool) for instance in instances] input_ids = [torch.cat([prompt[i], completion[i]], dim=-1) for i in range(len(instances))] loss_mask = [torch.cat([prompt_zeros[i], completion_ones[i]], dim=-1) for i in range(len(instances))] # right padding input_ids = torch.nn.utils.rnn.pad_sequence( sequences=input_ids, batch_first=True, padding_value=self.tokenizer.pad_token_id, ) # (bsz, max_len) loss_mask = torch.nn.utils.rnn.pad_sequence( sequences=loss_mask, batch_first=True, padding_value=0 ) # (bsz, max_len) to_pad = self.max_length - input_ids.size(1) input_ids = F.pad(input_ids, (0, to_pad), value=self.tokenizer.pad_token_id) loss_mask = F.pad(loss_mask, (0, to_pad), value=0) attention_mask = input_ids.ne(self.tokenizer.pad_token_id) # `torch.BoolTensor`, (bsz, max_len) # prepare kt data kl_completion = completion[::-1] # y' kl_completion_ones = [torch.ones_like(t) for t in kl_completion] kl_input_ids = [torch.cat([prompt[i], kl_completion[i]], dim=-1) for i in range(len(instances))] kl_loss_mask = [torch.cat([prompt_zeros[i], kl_completion_ones[i]], dim=-1) for i in range(len(instances))] # right padding kl_input_ids = torch.nn.utils.rnn.pad_sequence( sequences=kl_input_ids, batch_first=True, padding_value=self.tokenizer.pad_token_id, ) # (bsz, max_len) kl_loss_mask = torch.nn.utils.rnn.pad_sequence( sequences=kl_loss_mask, batch_first=True, padding_value=0 ) # (bsz, max_len) to_pad = self.max_length - kl_input_ids.size(1) kl_input_ids = F.pad(kl_input_ids, (0, to_pad), value=self.tokenizer.pad_token_id) kl_loss_mask = F.pad(kl_loss_mask, (0, to_pad), value=0) kl_attention_mask = kl_input_ids.ne(self.tokenizer.pad_token_id) # `torch.BoolTensor`, (bsz, max_len) data_dict = { "input_ids": input_ids, "attention_mask": attention_mask, "loss_mask": loss_mask, "label": torch.stack(label), "kl_input_ids": kl_input_ids, "kl_attention_mask": kl_attention_mask, "kl_loss_mask": kl_loss_mask, } return data_dict class StatefulDistributedSampler(DistributedSampler): def __init__( self, dataset: Dataset, num_replicas: Optional[int] = None, rank: Optional[int] = None, shuffle: bool = True, seed: int = 0, drop_last: bool = False, ) -> None: super().__init__(dataset, num_replicas, rank, shuffle, seed, drop_last) self.start_index: int = 0 def __iter__(self) -> Iterator: iterator = super().__iter__() indices = list(iterator) indices = indices[self.start_index :] return iter(indices) def __len__(self) -> int: return self.num_samples - self.start_index def set_start_index(self, start_index: int) -> None: self.start_index = start_index def apply_chat_template_and_mask( tokenizer: PreTrainedTokenizer, chat: List[Dict[str, str]], max_length: Optional[int] = None, system_prompt: str = None, padding: bool = True, truncation: bool = True, ignore_idx: int = -100, ) -> Dict[str, torch.Tensor]: if system_prompt is None: system_prompt = "You are a helpful assistant. The assistant first thinks about the reasoning process in the mind and then provides the user with the answer. The reasoning process and answer are enclosed within <think> </think> and<answer> </answer> tags, respectively, i.e., <think> reasoning process here </think><answer> answer here </answer>. Now the user asks you to solve a math problem that involves reasoning. After thinking, when you finally reach a conclusion, clearly output the final answer without explanation within the <answer> </answer> tags, i.e., <answer> 123 </answer>.\n\n" system_element = { "role": "system", "content": system_prompt, } # Format for RL. if "messages" in chat: gt_answer = chat.get("gt_answer", None) test_cases = chat.get("test_cases", None) chat = [chat["messages"]] tokens = [] assistant_mask = [] for i, msg in enumerate(chat): msg_tokens = tokenizer.apply_chat_template([system_element, msg], tokenize=True, add_generation_prompt=True) # remove unexpected bos token if i > 0 and msg_tokens[0] == tokenizer.bos_token_id: msg_tokens = msg_tokens[1:] tokens.extend(msg_tokens) if msg["role"] == "assistant": assistant_mask.extend([True] * len(msg_tokens)) else: assistant_mask.extend([False] * len(msg_tokens)) attention_mask = [1] * len(tokens) if max_length is not None: if padding and len(tokens) < max_length: to_pad = max_length - len(tokens) # Left padding for generation. tokens = [tokenizer.pad_token_id] * to_pad + tokens assistant_mask = [False] * to_pad + assistant_mask attention_mask = [0] * to_pad + attention_mask if truncation and len(tokens) > max_length: tokens = tokens[:max_length] assistant_mask = assistant_mask[:max_length] attention_mask = attention_mask[:max_length] input_ids = torch.tensor(tokens, dtype=torch.long) attention_mask = torch.tensor(attention_mask, dtype=torch.long) labels = input_ids.clone() labels[~torch.tensor(assistant_mask, dtype=torch.bool)] = ignore_idx if gt_answer is not None: return {"input_ids": input_ids, "attention_mask": attention_mask, "labels": labels, "gt_answer": gt_answer} elif test_cases is not None: return { "input_ids": input_ids, "attention_mask": attention_mask, "labels": labels, "test_cases": test_cases, } return { "input_ids": input_ids, "attention_mask": attention_mask, "labels": labels, } class RawConversationDataset(Dataset): """ Raw conversation dataset. Each instance is a dictionary with fields `system`, `roles`, `messages`, `offset`, `sep_style`, `seps`. """ def __init__(self, tokenizer: PreTrainedTokenizer, input_file: str, max_length: int, system_prompt: str) -> None: self.tokenizer = tokenizer self.raw_texts = [] with jsonlines.open(input_file) as f: for line in f: self.raw_texts.append(line) self.tokenized_texts = [None] * len(self.raw_texts) self.max_length = max_length self.system_prompt = system_prompt def __len__(self) -> int: return len(self.raw_texts) def __getitem__(self, index: int): if self.tokenized_texts[index] is None: message = self.raw_texts[index] tokens = apply_chat_template_and_mask(self.tokenizer, message, self.max_length, self.system_prompt) self.tokenized_texts[index] = dict(tokens) return self.tokenized_texts[index] def collate_fn_grpo(batch): input_ids = [item["input_ids"] for item in batch] attention_mask = [item["attention_mask"] for item in batch] labels = [item["labels"] for item in batch] # Assume input_ids, attention_mask, labels are already of the same length, # otherwise use pad_sequence(input_ids, batch_first=True, padding_value=tokenizer.pad_token_id) input_ids = torch.stack(input_ids) attention_mask = torch.stack(attention_mask) labels = torch.stack(labels) ret = {"input_ids": input_ids, "attention_mask": attention_mask, "labels": labels} if "test_cases" in batch[0]: ret["test_cases"] = [item["test_cases"] for item in batch] if "gt_answer" in batch[0]: ret["gt_answer"] = [item["gt_answer"] for item in batch] return ret
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/dataset/conversation.py
applications/ColossalChat/coati/dataset/conversation.py
import dataclasses import json import os from typing import Any, Dict, List import torch.distributed as dist from transformers import AutoTokenizer, PreTrainedTokenizer from colossalai.logging import get_dist_logger logger = get_dist_logger() @dataclasses.dataclass class Conversation: tokenizer: PreTrainedTokenizer system_message: str chat_template: str stop_ids: List[int] end_of_assistant: str roles = ["user", "assistant"] @classmethod def from_config(cls, tokenizer: PreTrainedTokenizer, config: Dict): """ Setup the conversation template from config """ tokenizer.chat_template = config["chat_template"] conv = cls( tokenizer, config["system_message"], config["chat_template"], config["stop_ids"], config["end_of_assistant"] ) conv.clear() return conv def clear(self): self.messages = [] @classmethod def get_conversation_template_keys(cls): return ["system_message", "chat_template"] def __str__(self): return json.dumps( {k: self.__dict__[k] for k in self.__dict__ if k not in ["tokenizer", "messages"]}, ensure_ascii=False, indent=4, ) def get_prompt(self, length: int = None, add_generation_prompt=False) -> Any: """ Retrieves the prompt for the conversation. Args: length (int, optional): The number of messages to include in the prompt. Defaults to None. get_seps_info (bool, optional): Whether to include separator information in the output. Defaults to False. add_generation_prompt (bool, optional): Whether to add the assistant line start token in generation (for generation only). Defaults to False. Returns: str or tuple: The prompt string if get_seps_info is False, otherwise a tuple containing the prompt string and separator information. """ if length is None: length = len(self.messages) assert length <= len(self.messages) if self.system_message is not None: messages = [{"role": "system", "content": self.system_message}] + self.messages[:length] else: messages = self.messages[:length] prompt = self.tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=add_generation_prompt ) return prompt def save_prompt(self): return self.get_prompt() def append_message(self, role: str, message: str): """ Append a message to the conversation. Args: role (str): The role of the message sender. Must be either 'user' or 'assistant'. message (str): The content of the message. Raises: AssertionError: If the role is not 'user' or 'assistant'. """ assert role in self.roles self.messages.append({"role": role, "content": message}) def copy(self): return Conversation(tokenizer=self.tokenizer, chat_template=self.chat_template) def setup_conversation_template( tokenizer: PreTrainedTokenizer, chat_template_config: Dict = None, save_path: str = None ) -> Conversation: """ Setup the conversation template, if chat_template is given, will replace the default chat_template of the tokenizer with it. Otherwise, the default chat_template will be used. If the tokenizer doesn't have a default chat_template, raise error to remind the user to set it manually. Args: tokenizer: The tokenizer to use chat_template_config: { "system_message": str The system message to use "chat_template": str The chat_template to use, if can be a chat_template, a huggingface model path or a local model. if a huggeface model path or a local model, the chat_template will be loaded from the model's tokenizer's default chat template. "stop_ids": List[int], the token ids used to terminate generation. You need to provide this for ppo training and generation. } """ if any([s not in chat_template_config.keys() for s in Conversation.get_conversation_template_keys()]): # Try to automatically set up conversation template, if fail, it throws an error that you need to do it manually if "end_of_assistant" not in chat_template_config: raise ValueError("Please set the end of assistant token.") if "system_message" not in chat_template_config: logger.warning("No system message is provided, will not use system message.") if "chat_template" not in chat_template_config: logger.warning("No chat_template is provided, will try to load it from the tokenizer.") if tokenizer.chat_template != None: chat_template_config["chat_template"] = tokenizer.chat_template else: raise ValueError( f"Load a tokenizer from {chat_template_config['chat_template']}, which doesn't have a default chat template, please set it manually." ) else: try: tokenizer = AutoTokenizer.from_pretrained(chat_template_config["chat_template"]) if tokenizer.chat_template != None: chat_template_config["chat_template"] = tokenizer.chat_template else: raise ValueError( f"Load a tokenizer from {chat_template_config['chat_template']}, which doesn't have a default chat template, please set it manually." ) logger.warning( f"chat_template is provided as a local model path or huggingface model path, loaded chat_template from \"{chat_template_config['chat_template']}\"." ) except OSError: pass except ValueError as e: raise ValueError(e) if save_path is not None and (not dist.is_initialized() or dist.get_rank() == 0): os.makedirs(os.path.dirname(save_path), exist_ok=True) with open(save_path, "w", encoding="utf8") as f: logger.info(f"Successfully generated a conversation tempalte config, save to {save_path}.") json.dump(chat_template_config, f, indent=4, ensure_ascii=False) return Conversation.from_config(tokenizer, chat_template_config)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/dataset/utils.py
applications/ColossalChat/coati/dataset/utils.py
import io import json from typing import Any, Dict, List import torch import torch.distributed as dist import torch.nn.functional as F from transformers import PreTrainedTokenizer def is_rank_0() -> bool: return not dist.is_initialized() or dist.get_rank() == 0 def _make_r_io_base(f, mode: str): if not isinstance(f, io.IOBase): f = open(f, mode=mode) return f def jload(f, mode="r"): """Load a .json file into a dictionary.""" f = _make_r_io_base(f, mode) jdict = json.load(f) f.close() return jdict def read_string_by_schema(data: Dict[str, Any], schema: str) -> str: """ Read a feild of the dataset be schema Args: data: Dict[str, Any] schema: cascaded feild names seperated by '.'. e.g. person.name.first will access data['person']['name']['first'] """ keys = schema.split(".") result = data for key in keys: result = result.get(key, None) if result is None: return "" assert isinstance(result, str), f"dataset element is not a string: {result}" return result def pad_to_max_len( sequence: List[torch.Tensor], max_length: int, padding_value: int, batch_first: bool = True, padding_side="left" ): """ Args: sequence: a batch of tensor of shape [batch_size, seq_len] if batch_first==True """ if padding_side == "left": reversed_sequence = [seq.flip(dims=(0,)) for seq in sequence] padded = torch.nn.utils.rnn.pad_sequence( sequences=reversed_sequence, batch_first=batch_first, padding_value=padding_value ) to_pad = max_length - padded.size(1) padded = F.pad(padded, (0, to_pad), value=padding_value) return torch.flip(padded, dims=(1,)) elif padding_side == "right": padded = torch.nn.utils.rnn.pad_sequence( sequences=sequence, batch_first=batch_first, padding_value=padding_value ) to_pad = max_length - padded.size(1) return F.pad(padded, (0, to_pad), value=padding_value) else: raise RuntimeError(f"`padding_side` can only be `left` or `right`, " f"but now `{padding_side}`") def chuncate_sequence(sequence: List[torch.Tensor], max_length: int, dtype: Any): """ Args: sequence: a batch of tensor of shape [batch_size, seq_len] if batch_first==True """ return [ torch.Tensor(seq[:max_length]).to(dtype) if len(seq) > max_length else torch.Tensor(seq).to(dtype) for seq in sequence ] def find_first_occurrence_subsequence(seq: torch.Tensor, subseq: torch.Tensor, start_index: int = 0) -> int: if subseq is None: return 0 for i in range(start_index, len(seq) - len(subseq) + 1): if torch.all(seq[i : i + len(subseq)] == subseq): return i return -1 def tokenize_and_concatenate( tokenizer: PreTrainedTokenizer, text: List[str], require_loss: List[bool], max_length: int, discard_non_loss_tokens_at_tail: bool = True, ): """ Tokenizes a list of texts using the provided tokenizer and concatenates the tokenized outputs. Args: tokenizer (PreTrainedTokenizer): The tokenizer to use for tokenization. text (List[str]): The list of texts to tokenize. require_loss (List[bool]): A list of boolean values indicating whether each text requires loss calculation. max_length: used to truncate the input ids discard_non_loss_tokens_at_tail: whether to discard the non-loss tokens at the tail if the first round has already exeeded max length - if the user query already exeeded max length, discard the sample - if only the first assistant response exeeded max length, truncate the response to fit the max length else keep the first several complete rounds of the conversations until max length is reached Returns: Tuple[List[int], List[int], List[int]]: A tuple containing the concatenated tokenized input ids, the start positions of loss spans, and the end positions of loss spans. """ input_ids = [] loss_starts = [] loss_ends = [] for s, r in zip(text, require_loss): tokenized = tokenizer(s, add_special_tokens=False)["input_ids"] if not max_length or len(input_ids) + len(tokenized) <= max_length or len(loss_ends) == 0: if r: loss_starts.append(len(input_ids)) loss_ends.append(len(input_ids) + len(tokenized)) input_ids.extend(tokenized) if max_length and loss_starts[0] >= max_length: return None, None, None if discard_non_loss_tokens_at_tail: input_ids = input_ids[: loss_ends[-1]] if max_length: input_ids = input_ids[:max_length] loss_ends[-1] = min(max_length, loss_ends[-1]) return input_ids, loss_starts, loss_ends def split_templated_prompt_into_chunks(messages: List[Dict[str, str]], prompt: str, end_of_assistant: str): # Seperate templated prompt into chunks by human/assistant's lines, prepare data for tokenize_and_concatenate start_idx = 0 chunks = [] require_loss = [] for line in messages: content_length = len(line["content"]) first_occur = prompt.find(line["content"], start_idx) if line["role"].lower() == "assistant" and end_of_assistant in prompt[first_occur + content_length :]: content_length = ( prompt.find(end_of_assistant, first_occur + content_length) + len(end_of_assistant) - first_occur ) # if the tokenized content start with a leading space, we want to keep it in loss calculation # e.g., Assistant: I am saying... # if the tokenized content doesn't start with a leading space, we only need to keep the content in loss calculation # e.g., # Assistant: # '\n' as line breaker # I am saying... if prompt[first_occur - 1] != " ": chunks.append(prompt[start_idx:first_occur]) chunks.append(prompt[first_occur : first_occur + content_length]) else: chunks.append(prompt[start_idx : first_occur - 1]) chunks.append(prompt[first_occur - 1 : first_occur + content_length]) start_idx = first_occur + content_length if line["role"].lower() == "assistant": require_loss.append(False) require_loss.append(True) else: require_loss.append(False) require_loss.append(False) chunks.append(prompt[start_idx:]) require_loss.append(False) return chunks, require_loss
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/dataset/tokenization_utils.py
applications/ColossalChat/coati/dataset/tokenization_utils.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ tokenization utils for constructing dataset for ppo, dpo, sft, rm """ import warnings from copy import deepcopy from typing import Any, Dict, List, Union from coati.dataset.conversation import Conversation from coati.dataset.utils import split_templated_prompt_into_chunks, tokenize_and_concatenate from datasets import dataset_dict from torch.utils.data import ConcatDataset, Dataset from transformers import PreTrainedTokenizer from colossalai.logging import get_dist_logger logger = get_dist_logger() IGNORE_INDEX = -100 DSType = Union[Dataset, ConcatDataset, dataset_dict.Dataset] def tokenize_sft( data_point: Dict[str, str], tokenizer: PreTrainedTokenizer, conversation_template: Conversation = None, max_length: int = 4096, ) -> Dict[str, Union[int, str, List[int]]]: """ A tokenization function to tokenize an original pretraining data point as following and calculate corresponding labels for sft training: "Something here can be system message[user_line_start]User line[User line end][Assistant line start]Assistant line[Assistant line end]...[Assistant line end]Something here" ^ end_of_system_line_position Args: data_point: the data point of the following format {"messages": [{"from": "user", "content": "xxx"}, {"from": "assistant", "content": "xxx"}]} tokenizer: the tokenizer whose conversation_template: the conversation template to apply ignore_index: the ignore index when calculate loss during training max_length: the maximum context length """ ignore_index = IGNORE_INDEX messages = data_point["messages"] template = deepcopy(conversation_template) if messages[0]["from"] == "system": template.system_message = str(messages[0]["content"]) messages.pop(0) template.messages = [] for idx, mess in enumerate(messages): if mess["from"] != template.roles[idx % 2]: raise ValueError( f"Message should iterate between user and assistant and starts with a \ line from the user. Got the following data:\n{messages}" ) template.append_message(mess["from"], mess["content"]) if len(template.messages) % 2 != 0: # Force to end with assistant response template.messages = template.messages[0:-1] # tokenize and calculate masked labels -100 for positions corresponding to non-assistant lines prompt = template.get_prompt() chunks, require_loss = split_templated_prompt_into_chunks( template.messages, prompt, conversation_template.end_of_assistant ) tokenized, starts, ends = tokenize_and_concatenate(tokenizer, chunks, require_loss, max_length=max_length) if tokenized is None: return dict( input_ids=None, labels=None, inputs_decode=None, labels_decode=None, seq_length=None, seq_category=None, ) labels = [ignore_index] * len(tokenized) for start, end in zip(starts, ends): labels[start:end] = tokenized[start:end] if tokenizer.bos_token_id is not None: # Force to add bos token at the beginning of the tokenized sequence if the input ids doesn;t starts with bos if tokenized[0] != tokenizer.bos_token_id: # Some chat templates already include bos token tokenized = [tokenizer.bos_token_id] + tokenized labels = [-100] + labels # log decoded inputs and labels for debugging inputs_decode = tokenizer.decode(tokenized) start = 0 end = 0 label_decode = [] for i in range(len(labels)): if labels[i] == ignore_index: if start != end: label_decode.append(tokenizer.decode(labels[start + 1 : i], skip_special_tokens=False)) start = i end = i else: end = i if i == len(labels) - 1: label_decode.append(tokenizer.decode(labels[start + 1 :], skip_special_tokens=False)) # Check if all labels are ignored, this may happen when the tokenized length is too long if labels.count(ignore_index) == len(labels): return dict( input_ids=None, labels=None, inputs_decode=None, labels_decode=None, seq_length=None, seq_category=None, ) return dict( input_ids=tokenized, labels=labels, inputs_decode=inputs_decode, labels_decode=label_decode, seq_length=len(tokenized), seq_category=data_point["category"] if "category" in data_point else "None", ) def tokenize_prompt( data_point: Dict[str, str], tokenizer: PreTrainedTokenizer, conversation_template: Conversation = None, max_length: int = 4096, ) -> Dict[str, Union[int, str, List[int]]]: """ A tokenization function to tokenize an original pretraining data point as following for ppo training: "Something here can be system message[user_line_start]User line[User line end][Assistant line start]Assistant line[Assistant line end]...[Assistant line start]" Args: data_point: the data point of the following format {"messages": [{"from": "user", "content": "xxx"}, {"from": "assistant", "content": "xxx"}]} tokenizer: the tokenizer whose conversation_template: the conversation template to apply ignore_index: the ignore index when calculate loss during training max_length: the maximum context length """ messages = data_point["messages"] template = deepcopy(conversation_template) template.messages = [] if messages[0]["from"] == "system": template.system_message = str(messages[0]["content"]) messages.pop(0) for idx, mess in enumerate(messages): if mess["from"] != template.roles[idx % 2]: raise ValueError( f"Message should iterate between user and assistant and starts with a line from the user. Got the following data:\n{messages}" ) template.append_message(mess["from"], mess["content"]) # `target_turn_index` is the number of turns which exceeds `max_length - 1` for the first time. if len(template.messages) % 2 != 1: # exclude the answer if provided. keep only the prompt template.messages = template.messages[:-1] # Prepare data prompt = template.get_prompt(length=len(template.messages), add_generation_prompt=True) tokenized = tokenizer([prompt], add_special_tokens=False)["input_ids"][0] if tokenizer.bos_token_id is not None: if tokenized[0] != tokenizer.bos_token_id: tokenized = [tokenizer.bos_token_id] + tokenized if len(tokenized) > max_length: return dict( input_ids=None, inputs_decode=None, seq_length=None, seq_category=None, ) # `inputs_decode` can be used to check whether the tokenization method is true. if "gt_answer" in data_point: return dict( input_ids=tokenized, inputs_decode=prompt, seq_length=len(tokenized), seq_category=data_point["category"] if "category" in data_point else "None", gt_answer=data_point["gt_answer"], ) else: return dict( input_ids=tokenized, inputs_decode=prompt, seq_length=len(tokenized), seq_category=data_point["category"] if "category" in data_point else "None", ) def apply_rlhf_data_format(template: Conversation, tokenizer: Any): target_turn = int(len(template.messages) / 2) prompt = template.get_prompt(target_turn * 2) chunks, require_loss = split_templated_prompt_into_chunks( template.messages[: 2 * target_turn], prompt, template.end_of_assistant ) # no truncation applied tokenized, starts, ends = tokenize_and_concatenate(tokenizer, chunks, require_loss, max_length=None) loss_mask = [0] * len(tokenized) label_decode = [] # only the last round (chosen/rejected) is used to calculate loss for i in range(starts[-1], ends[-1]): loss_mask[i] = 1 label_decode.append(tokenizer.decode(tokenized[starts[-1] : ends[-1]], skip_special_tokens=False)) if tokenizer.bos_token_id is not None: if tokenized[0] != tokenizer.bos_token_id: tokenized = [tokenizer.bos_token_id] + tokenized loss_mask = [0] + loss_mask return {"input_ids": tokenized, "loss_mask": loss_mask, "label_decode": label_decode} def tokenize_rlhf( data_point: Dict[str, str], tokenizer: PreTrainedTokenizer, conversation_template: Conversation = None, max_length: int = 4096, ) -> Dict[str, Union[int, str, List[int]]]: """ A tokenization function to tokenize an original pretraining data point as following: {"context": [{"from": "user", "content": "xxx"}, {"from": "assistant", "content": "xxx"}], "chosen": {"from": "assistant", "content": "xxx"}, "rejected": {"from": "assistant", "content": "xxx"}} """ context = data_point["context"] template = deepcopy(conversation_template) template.clear() if context[0]["from"] == "system": template.system_message = str(context[0]["content"]) context.pop(0) for idx, mess in enumerate(context): if mess["from"] != template.roles[idx % 2]: raise ValueError( f"Message should iterate between user and assistant and starts with a \ line from the user. Got the following data:\n{context}" ) template.append_message(mess["from"], mess["content"]) if len(template.messages) % 2 != 1: warnings.warn( "Please make sure leading context starts and ends with a line from user\nLeading context: " + str(template.messages) ) return dict( chosen_input_ids=None, chosen_loss_mask=None, chosen_label_decode=None, rejected_input_ids=None, rejected_loss_mask=None, rejected_label_decode=None, ) assert context[-1]["from"].lower() == template.roles[0], "The last message in context should be from user." chosen = deepcopy(template) rejected = deepcopy(template) chosen_continuation = data_point["chosen"] rejected_continuation = data_point["rejected"] for round in range(len(chosen_continuation)): if chosen_continuation[round]["from"] != template.roles[(round + 1) % 2]: raise ValueError( f"Message should iterate between user and assistant and starts with a \ line from the user. Got the following data:\n{chosen_continuation}" ) chosen.append_message(chosen_continuation[round]["from"], chosen_continuation[round]["content"]) for round in range(len(rejected_continuation)): if rejected_continuation[round]["from"] != template.roles[(round + 1) % 2]: raise ValueError( f"Message should iterate between user and assistant and starts with a \ line from the user. Got the following data:\n{rejected_continuation}" ) rejected.append_message(rejected_continuation[round]["from"], rejected_continuation[round]["content"]) ( chosen_input_ids, chosen_loss_mask, chosen_label_decode, rejected_input_ids, rejected_loss_mask, rejected_label_decode, ) = (None, None, None, None, None, None) chosen_data_packed = apply_rlhf_data_format(chosen, tokenizer) (chosen_input_ids, chosen_loss_mask, chosen_label_decode) = ( chosen_data_packed["input_ids"], chosen_data_packed["loss_mask"], chosen_data_packed["label_decode"], ) rejected_data_packed = apply_rlhf_data_format(rejected, tokenizer) (rejected_input_ids, rejected_loss_mask, rejected_label_decode) = ( rejected_data_packed["input_ids"], rejected_data_packed["loss_mask"], rejected_data_packed["label_decode"], ) if len(chosen_input_ids) > max_length or len(rejected_input_ids) > max_length: return dict( chosen_input_ids=None, chosen_loss_mask=None, chosen_label_decode=None, rejected_input_ids=None, rejected_loss_mask=None, rejected_label_decode=None, ) # Check if loss mask is all 0s (no loss), this may happen when the tokenized length is too long if chosen_loss_mask.count(1) == 0 or rejected_loss_mask.count(1) == 0: return dict( chosen_input_ids=None, chosen_loss_mask=None, chosen_label_decode=None, rejected_input_ids=None, rejected_loss_mask=None, rejected_label_decode=None, ) return { "chosen_input_ids": chosen_input_ids, "chosen_loss_mask": chosen_loss_mask, "chosen_label_decode": chosen_label_decode, "rejected_input_ids": rejected_input_ids, "rejected_loss_mask": rejected_loss_mask, "rejected_label_decode": rejected_label_decode, } def tokenize_kto( data_point: Dict[str, str], tokenizer: PreTrainedTokenizer, conversation_template: Conversation = None, max_length: int = 4096, ) -> Dict[str, Union[int, str, List[int]]]: """ Tokenize a dataset for KTO training The raw input data is conversation that have the following format { "prompt": [{"from": "user", "content": "xxx"}...], "completion": {"from": "assistant", "content": "xxx"}, "label": true/false } It returns three fields The context, which contain the query and the assistant start, the completion, which only contains the assistance's answer, and a binary label, which indicates if the sample is prefered or not """ prompt = data_point["prompt"] completion = data_point["completion"] template = deepcopy(conversation_template) template.clear() if prompt[0]["from"] == "system": template.system_message = str(prompt[0]["content"]) prompt.pop(0) if prompt[0].get("from", None) != "user": raise ValueError("conversation should start with user") if completion.get("from", None) != "assistant": raise ValueError("conversation should end with assistant") for mess in prompt: if mess.get("from", None) == "user": template.append_message("user", mess["content"]) elif mess.get("from", None) == "assistant": template.append_message("assistant", mess["content"]) else: raise ValueError(f"Unsupported role {mess.get('from', None)}") generation_prompt = template.get_prompt(len(prompt), add_generation_prompt=True) template.append_message("assistant", completion["content"]) full_prompt = template.get_prompt(len(prompt) + 1, add_generation_prompt=False) tokenized_full_prompt = tokenizer(full_prompt, add_special_tokens=False)["input_ids"] if len(tokenized_full_prompt) + 1 > max_length: return dict(prompt=None, completion=None, label=None, input_id_decode=None, completion_decode=None) tokenized_generation_prompt = tokenizer(generation_prompt, add_special_tokens=False)["input_ids"] tokenized_completion = tokenized_full_prompt[len(tokenized_generation_prompt) :] tokenized_completion = deepcopy(tokenized_completion) if tokenizer.bos_token_id is not None and tokenized_generation_prompt[0] != tokenizer.bos_token_id: tokenized_generation_prompt = [tokenizer.bos_token_id] + tokenized_generation_prompt decoded_full_prompt = tokenizer.decode(tokenized_full_prompt, skip_special_tokens=False) decoded_completion = tokenizer.decode(tokenized_completion, skip_special_tokens=False) return { "prompt": tokenized_generation_prompt, "completion": tokenized_completion, "label": data_point["label"], "input_id_decode": decoded_full_prompt, "completion_decode": decoded_completion, }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/dataset/__init__.py
applications/ColossalChat/coati/dataset/__init__.py
from .conversation import Conversation, setup_conversation_template from .loader import ( DataCollatorForKTODataset, DataCollatorForPreferenceDataset, DataCollatorForPromptDataset, DataCollatorForSupervisedDataset, StatefulDistributedSampler, load_tokenized_dataset, ) from .tokenization_utils import tokenize_kto, tokenize_prompt, tokenize_rlhf, tokenize_sft __all__ = [ "tokenize_prompt", "DataCollatorForPromptDataset", "is_rank_0", "DataCollatorForPreferenceDataset", "DataCollatorForSupervisedDataset", "DataCollatorForKTODataset", "StatefulDistributedSampler", "load_tokenized_dataset", "tokenize_sft", "tokenize_rlhf", "tokenize_kto", "setup_conversation_template", "Conversation", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/detached_trainer_base.py
applications/ColossalChat/coati/ray/detached_trainer_base.py
import os from abc import ABC, abstractmethod from typing import Any, Dict, List import ray import torch from coati.experience_buffer.utils import BufferItem from coati.experience_maker import Experience from torch.utils.data import DataLoader from tqdm import tqdm from .callbacks import TrainerCallback from .detached_replay_buffer import DetachedReplayBuffer from .utils import is_rank_0 class DetachedTrainer(ABC): """ Base class for detached rlhf trainers. 'detach' means that the experience maker is detached compared to a normal Trainer. Please set name attribute during init: >>> trainer = DetachedTrainer.options(..., name = "xxx", ...).remote() So an ExperienceMakerHolder can reach the detached_replay_buffer by Actor's name. Args: detached_strategy (DetachedStrategy): the strategy to use for training detached_replay_buffer_ref (ObjectRef[DetachedReplayBuffer]): the replay buffer to use for training data_loader_pin_memory (bool, defaults to True): whether to pin memory for data loader callbacks (List[Callback], defaults to []): the callbacks to call during training process generate_kwargs (dict, optional): the kwargs to use while model generating """ def __init__( self, experience_maker_holder_name_list: List[str], train_batch_size: int = 8, buffer_limit: int = 0, dataloader_pin_memory: bool = True, callbacks: List[TrainerCallback] = [], debug: bool = False, ) -> None: super().__init__() self.detached_replay_buffer = DetachedReplayBuffer(train_batch_size, limit=buffer_limit) self.dataloader_pin_memory = dataloader_pin_memory self.callbacks = callbacks self.target_holder_name_list = experience_maker_holder_name_list self.target_holder_list = [] self._is_target_holder_initialized = False self._debug = debug def update_target_holder_list(self): # as the length of target_holder_list may be zero, we need to check it by a bool flag if not self._is_target_holder_initialized: for name in self.target_holder_name_list: self.target_holder_list.append(ray.get_actor(name, namespace=os.environ["RAY_NAMESPACE"])) self._is_target_holder_initialized = True @abstractmethod def _update_remote_makers(self, fully_update: bool = False, **kwargs): pass def sync_models_to_remote_makers(self, **kwargs): self._update_remote_makers(fully_update=True, **kwargs) @abstractmethod def training_step(self, experience: Experience) -> Dict[str, Any]: pass def _learn(self, update_steps: int, train_epochs: int) -> None: data = [] # warmup pbar = tqdm(range(update_steps), desc=f"Train epoch [1/{train_epochs}]", disable=not is_rank_0()) self._on_epoch_start(0) self._learn_epoch(pbar, data) self._on_epoch_end(0) # item is already a batch dataloader = DataLoader( data, batch_size=1, shuffle=True, pin_memory=self.dataloader_pin_memory, collate_fn=lambda x: x[0] ) for epoch in range(1, train_epochs): pbar = tqdm(dataloader, desc=f"Train epoch [{epoch + 1}/{train_epochs}]", disable=not is_rank_0()) self._on_epoch_start(epoch) self._learn_epoch(pbar, data) self._on_epoch_end(epoch) def _learn_epoch(self, pbar: tqdm, data: List[Experience]) -> None: is_warmup = len(data) == 0 for x in pbar: if self._debug: print("[trainer] training step") # sample a batch and then train to avoid waiting experience = x if not is_warmup else self._buffer_sample() experience.to_device(torch.cuda.current_device()) self._on_batch_start() metrics = self.training_step(experience) self._on_batch_end(metrics, experience) if self._debug: print("[trainer] step over") experience.to_device("cpu") if is_warmup: data.append(experience) pbar.set_postfix(metrics) def fit(self, total_steps: int, update_steps: int, train_epochs: int = 1) -> None: self._on_fit_start() for i in tqdm(range(total_steps // update_steps), desc="Trainer", disable=not is_rank_0()): self._on_episode_start(i) self._learn(update_steps, train_epochs) self._on_update_start() self._update_remote_makers() self._on_update_end() self._on_episode_end(i) self._on_fit_end() @ray.method(concurrency_group="buffer_length") def buffer_get_length(self): # called by ExperienceMakerHolder if self._debug: print("[trainer] telling length") return self.detached_replay_buffer.get_length() @ray.method(concurrency_group="buffer_append") def buffer_append(self, experience: Experience): # called by ExperienceMakerHolder if self._debug: print(f"[trainer] receiving exp.") self.detached_replay_buffer.append(experience) @ray.method(concurrency_group="buffer_append") def buffer_extend(self, items: List[BufferItem]): # called by ExperienceMakerHolder if self._debug: print(f"[trainer] receiving exp.") self.detached_replay_buffer.extend(items) @ray.method(concurrency_group="buffer_sample") def _buffer_sample(self): return self.detached_replay_buffer.sample() def _on_fit_start(self) -> None: for callback in self.callbacks: callback.on_fit_start() def _on_fit_end(self) -> None: for callback in self.callbacks: callback.on_fit_end() def _on_episode_start(self, episode: int) -> None: for callback in self.callbacks: callback.on_episode_start(episode) def _on_episode_end(self, episode: int) -> None: for callback in self.callbacks: callback.on_episode_end(episode) def _on_epoch_start(self, epoch: int) -> None: for callback in self.callbacks: callback.on_epoch_start(epoch) def _on_epoch_end(self, epoch: int) -> None: for callback in self.callbacks: callback.on_epoch_end(epoch) def _on_batch_start(self) -> None: for callback in self.callbacks: callback.on_batch_start() def _on_batch_end(self, metrics: dict, experience: Experience) -> None: for callback in self.callbacks: callback.on_batch_end(metrics, experience) def _on_update_start(self) -> None: for callback in self.callbacks: callback.on_update_start() def _on_update_end(self) -> None: for callback in self.callbacks: callback.on_update_end()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/utils.py
applications/ColossalChat/coati/ray/utils.py
import os from collections import OrderedDict from typing import Any, Dict import torch import torch.distributed as dist import torch.nn as nn from coati.models.bloom import BLOOMRM, BLOOMActor, BLOOMCritic from coati.models.gpt import GPTRM, GPTActor, GPTCritic from coati.models.llama import LlamaActor, LlamaCritic, LlamaRM from coati.models.opt import OPTRM, OPTActor, OPTCritic from coati.trainer.strategies import DDPStrategy, GeminiStrategy, LowLevelZeroStrategy from transformers import AutoTokenizer, BloomTokenizerFast, GPT2Tokenizer def is_rank_0() -> bool: return not dist.is_initialized() or dist.get_rank() == 0 def get_rank() -> int: return dist.get_rank() if dist.is_initialized() else 0 def get_world_size() -> int: return dist.get_world_size() if dist.is_initialized() else 1 def get_actor_from_args(model: str, pretrained: str = None, config=None, lora_rank=0): if model == "gpt2": actor = GPTActor(pretrained=pretrained, config=config, lora_rank=lora_rank) elif model == "bloom": actor = BLOOMActor(pretrained=pretrained, config=config, lora_rank=lora_rank) elif model == "opt": actor = OPTActor(pretrained=pretrained, config=config, lora_rank=lora_rank) elif model == "llama": actor = LlamaActor(pretrained=pretrained, config=config, lora_rank=lora_rank) else: raise ValueError(f'Unsupported actor model "{model}"') return actor def get_critic_from_args(model: str, pretrained: str = None, config=None, lora_rank=0): if model == "gpt2": critic = GPTCritic(pretrained=pretrained, lora_rank=lora_rank, config=config) elif model == "bloom": critic = BLOOMCritic(pretrained=pretrained, lora_rank=lora_rank, config=config) elif model == "opt": critic = OPTCritic(pretrained=pretrained, lora_rank=lora_rank, config=config) elif model == "llama": critic = LlamaCritic(pretrained=pretrained, lora_rank=lora_rank, config=config) else: raise ValueError(f'Unsupported reward model "{model}"') return critic def get_reward_model_from_args(model: str, pretrained: str = None, config=None): if model == "gpt2": reward_model = GPTRM(pretrained=pretrained, config=config) elif model == "bloom": reward_model = BLOOMRM(pretrained=pretrained, config=config) elif model == "opt": reward_model = OPTRM(pretrained=pretrained, config=config) elif model == "llama": reward_model = LlamaRM(pretrained=pretrained, config=config) else: raise ValueError(f'Unsupported reward model "{model}"') return reward_model def get_strategy_from_args(strategy: str): if strategy == "ddp": strategy_ = DDPStrategy() elif strategy == "colossalai_gemini": strategy_ = GeminiStrategy(placement_policy="static", initial_scale=2**5) elif strategy == "colossalai_zero2": strategy_ = LowLevelZeroStrategy(stage=2, placement_policy="cuda") elif strategy == "colossalai_gemini_cpu": strategy_ = GeminiStrategy( placement_policy="static", offload_optim_frac=1.0, offload_param_frac=1.0, initial_scale=2**5 ) elif strategy == "colossalai_zero2_cpu": strategy_ = LowLevelZeroStrategy(stage=2, placement_policy="cpu") else: raise ValueError(f'Unsupported strategy "{strategy}"') return strategy_ def get_tokenizer_from_args(model: str, **kwargs): if model == "gpt2": tokenizer = GPT2Tokenizer.from_pretrained("gpt2") elif model == "bloom": tokenizer = BloomTokenizerFast.from_pretrained("bigscience/bloom-560m") elif model == "opt": tokenizer = AutoTokenizer.from_pretrained("facebook/opt-350m") elif model == "llama": pretrain_path = kwargs["pretrain"] tokenizer = AutoTokenizer.from_pretrained(pretrain_path) else: raise ValueError(f'Unsupported model "{model}"') tokenizer.pad_token = tokenizer.eos_token return tokenizer def set_dist_env(env_info: Dict[str, str]): os.environ["RANK"] = env_info["rank"] os.environ["LOCAL_RANK"] = env_info["local_rank"] os.environ["WORLD_SIZE"] = env_info["world_size"] os.environ["MASTER_PORT"] = env_info["master_port"] os.environ["MASTER_ADDR"] = env_info["master_addr"] def get_model_numel(model: nn.Module) -> int: numel = sum(p.numel() for p in model.parameters()) return numel def get_receivers_per_sender(sender_idx: int, num_senders: int, num_receivers: int, allow_idle_sender: bool) -> list: target_receivers = [] if num_senders <= num_receivers or allow_idle_sender: # a sender will send data to one or more receivers # a receiver only has one sender for i in range(num_receivers): if i % num_senders == sender_idx: target_receivers.append(i) else: # a sender will send data to one receiver # a receiver may have more than one sender target_receivers.append(sender_idx % num_receivers) return target_receivers def state_dict_to( state_dict: Dict[str, Any], dtype: torch.dtype = torch.float16, device: torch.device = torch.device("cpu") ): """ keep state_dict intact """ new_state_dict = OrderedDict() for k, v in state_dict.items(): new_state_dict[k] = v.to(dtype=dtype, device=device) return new_state_dict
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/detached_trainer_ppo.py
applications/ColossalChat/coati/ray/detached_trainer_ppo.py
from typing import Callable, Dict, List, Tuple import ray import torch from coati.experience_maker import Experience from coati.models.base import Actor, Critic from coati.models.loss import PolicyLoss, ValueLoss from coati.trainer.strategies import GeminiStrategy, LowLevelZeroStrategy, Strategy from torch.optim import Adam from colossalai.nn.optimizer import HybridAdam from .callbacks import TrainerCallback, TrainerPerformanceEvaluator from .detached_trainer_base import DetachedTrainer from .lora_constructor import LoRAConstructor from .utils import get_model_numel, get_rank, set_dist_env, state_dict_to @ray.remote( concurrency_groups={"buffer_length": 1, "buffer_append": 1, "buffer_sample": 1, "model_io": 1, "compute": 1} ) class DetachedPPOTrainer(DetachedTrainer): """ Detached Trainer for PPO algorithm Args: strategy (Strategy): the strategy to use for training model (str) : for actor / critic init pretrained (str) : for actor / critic init lora_rank (int) : for actor / critic init train_batch_size (int, defaults to 8): the batch size to use for training train_batch_size (int, defaults to 8): the batch size to use for training buffer_limit (int, defaults to 0): the max_size limitation of replay buffer buffer_cpu_offload (bool, defaults to True): whether to offload replay buffer to cpu eps_clip (float, defaults to 0.2): the clip coefficient of policy loss value_clip (float, defaults to 0.4): the clip coefficient of value loss experience_batch_size (int, defaults to 8): the batch size to use for experience generation max_epochs (int, defaults to 1): the number of epochs of training process dataloader_pin_memory (bool, defaults to True): whether to pin memory for data loader callbacks (List[Callback], defaults to []): the callbacks to call during training process generate_kwargs (dict, optional): the kwargs to use while model generating """ def __init__( self, experience_maker_holder_name_list: List[str], strategy_fn: Callable[[], Strategy], model_fn: Callable[[], Tuple[Actor, Critic]], env_info: Dict[str, str] = None, train_batch_size: int = 8, buffer_limit: int = 0, eps_clip: float = 0.2, value_clip: float = 0.4, dataloader_pin_memory: bool = True, callbacks: List[TrainerCallback] = [], eval_performance: bool = False, debug: bool = False, update_lora_weights: bool = False, ) -> None: # set environment variables if env_info: set_dist_env(env_info=env_info) # configure strategy self.strategy = strategy_fn() # configure models, loss and optimizers with self.strategy.model_init_context(): self.actor, self.critic = model_fn() if eval_performance: actor_numel = get_model_numel(self.actor) critic_numel = get_model_numel(self.critic) evaluator = TrainerPerformanceEvaluator(actor_numel, critic_numel) callbacks = callbacks + [evaluator] if isinstance(self.strategy, (LowLevelZeroStrategy, GeminiStrategy)): self.actor_optim = HybridAdam(self.actor.parameters(), lr=1e-7) self.critic_optim = HybridAdam(self.critic.parameters(), lr=1e-7) else: self.actor_optim = Adam(self.actor.parameters(), lr=1e-7) self.critic_optim = Adam(self.critic.parameters(), lr=1e-7) (self.actor, self.actor_optim), (self.critic, self.critic_optim) = self.strategy.prepare( (self.actor, self.actor_optim), (self.critic, self.critic_optim) ) # configure trainer self.actor_loss_fn = PolicyLoss(eps_clip) self.critic_loss_fn = ValueLoss(value_clip) super().__init__( experience_maker_holder_name_list, train_batch_size=train_batch_size, buffer_limit=buffer_limit, dataloader_pin_memory=dataloader_pin_memory, callbacks=callbacks, debug=debug, ) if self._debug: print(f"[trainer{get_rank()}] will send state dict to {experience_maker_holder_name_list}") self._update_lora_weights = update_lora_weights @ray.method(concurrency_group="model_io") @torch.no_grad() def _update_remote_makers(self, fully_update: bool = False, **config): # TODO: balance duties if not fully_update: config["requires_grad_only"] = True self.update_target_holder_list() # mark start, ensure order tasks = [] for target_holder in self.target_holder_list: tasks.append(target_holder.update_experience_maker.remote(chunk_start=True, fully_update=fully_update)) ray.get(tasks) # sending loop tasks = [] for state_dict_shard in self._get_model_state_dict_shard(self.actor, fully_update=fully_update, **config): for target_holder in self.target_holder_list: tasks.append( target_holder.update_experience_maker.remote( new_actor_state_dict=state_dict_shard, new_actor_lora_config_dict=self._get_model_lora_config_dict(self.actor), fully_update=fully_update, ) ) # sending loop for state_dict_shard in self._get_model_state_dict_shard(self.critic, fully_update=fully_update, **config): for target_holder in self.target_holder_list: tasks.append( target_holder.update_experience_maker.remote( new_critic_state_dict=state_dict_shard, new_critic_lora_config_dict=self._get_model_lora_config_dict(self.critic), fully_update=fully_update, ) ) ray.get(tasks) # mark end for target_holder in self.target_holder_list: target_holder.update_experience_maker.remote(chunk_end=True, fully_update=fully_update) @ray.method(concurrency_group="compute") def training_step(self, experience: Experience) -> Dict[str, float]: self.actor.train() self.critic.train() num_actions = experience.action_mask.size(1) action_log_probs = self.actor(experience.sequences, num_actions, attention_mask=experience.attention_mask) actor_loss = self.actor_loss_fn( action_log_probs, experience.action_log_probs, experience.advantages, action_mask=experience.action_mask ) self.strategy.backward(actor_loss, self.actor, self.actor_optim) self.strategy.optimizer_step(self.actor_optim) self.actor_optim.zero_grad() values = self.critic( experience.sequences, action_mask=experience.action_mask, attention_mask=experience.attention_mask ) critic_loss = self.critic_loss_fn( values, experience.values, experience.reward, action_mask=experience.action_mask ) self.strategy.backward(critic_loss, self.critic, self.critic_optim) self.strategy.optimizer_step(self.critic_optim) self.critic_optim.zero_grad() return {"actor_loss": actor_loss.item(), "critic_loss": critic_loss.item()} def strategy_save_actor(self, path: str, only_rank0: bool = False) -> None: self.strategy.save_model(self.actor, path, only_rank0) def strategy_save_critic(self, path: str, only_rank0: bool = False) -> None: self.strategy.save_model(self.critic, path, only_rank0) def strategy_save_actor_optim(self, path: str, only_rank0: bool = False) -> None: self.strategy.save_optimizer(self.actor_optim, path, only_rank0) def strategy_save_critic_optim(self, path: str, only_rank0: bool = False) -> None: self.strategy.save_optimizer(self.critic_optim, path, only_rank0) def _get_model_state_dict_shard(self, model: torch.nn.Module, fully_update=False, **config): for state_dict in self.strategy.get_model_state_dict_shard(model, **config): if not self._update_lora_weights or fully_update: yield state_dict_to(state_dict) else: state_dict_lora, _ = LoRAConstructor.filter_state_dict_lora(state_dict) yield state_dict_to(state_dict_lora) def _get_model_lora_config_dict(self, model: torch.nn.Module): if not self._update_lora_weights: return None unwrapped_model = self.strategy.unwrap_model(model) return LoRAConstructor.extract_lora_config(unwrapped_model)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/experience_maker_holder.py
applications/ColossalChat/coati/ray/experience_maker_holder.py
import os import time import tracemalloc from threading import Lock from typing import Any, Callable, Dict, Iterable, List, Tuple, Union import ray import torch from coati.experience_buffer.utils import split_experience_batch from coati.experience_maker import Experience, NaiveExperienceMaker from coati.models.base import Actor, Critic, RewardModel from coati.trainer.strategies import Strategy from torch import Tensor from tqdm import tqdm from .callbacks import ExperienceMakerPerformanceEvaluator, MakerCallback from .lora_constructor import LoRAConstructor from .utils import get_model_numel, get_rank, is_rank_0, set_dist_env, state_dict_to @ray.remote(concurrency_groups={"experience_io": 1, "model_io": 1, "compute": 1}) class ExperienceMakerHolder: """ Args: detached_trainer_name_list: str list to get ray actor handles strategy: kl_coef: the coefficient of kl divergence loss sync_models_from_trainers: whether to sync models from trainers. If True, you must call sync_models_to_remote_makers() in trainers to sync models. """ def __init__( self, detached_trainer_name_list: List[str], strategy_fn: Callable[[], Strategy], # a function returns (actor, critic, reward_model, initial_model) model_fn: Callable[[], Tuple[Actor, Critic, RewardModel, Actor]], env_info: Dict[str, str] = None, sync_models_from_trainers: bool = False, buffer_cpu_offload: bool = True, kl_coef: float = 0.1, callbacks: List[MakerCallback] = [], eval_performance: bool = False, debug: bool = False, update_lora_weights: bool = False, **generate_kwargs, ): # set environment variables if env_info: set_dist_env(env_info=env_info) self.target_trainer_list = [] assert len(detached_trainer_name_list) > 0 self._detached_trainer_name_list = detached_trainer_name_list self.strategy = strategy_fn() self.buffer_cpu_offload = buffer_cpu_offload self.kl_coef = kl_coef # init models with self.strategy.model_init_context(): actor, critic, reward_model, initial_model = model_fn() self.generate_kwargs = _set_default_generate_kwargs(generate_kwargs, actor) if eval_performance: actor_numel = get_model_numel(actor) critic_numel = get_model_numel(critic) initial_model_numel = get_model_numel(initial_model) reward_model_numel = get_model_numel(reward_model) evaluator = ExperienceMakerPerformanceEvaluator( actor_numel, critic_numel, initial_model_numel, reward_model_numel ) callbacks = callbacks + [evaluator] actor, critic, reward_model, initial_model = self.strategy.prepare(actor, critic, reward_model, initial_model) self.experience_maker = NaiveExperienceMaker(actor, critic, reward_model, initial_model, self.kl_coef) self.callbacks = callbacks self._model_visit_lock = Lock() self._is_fully_initialized = not sync_models_from_trainers self._debug = debug self._update_lora_weights = update_lora_weights if self._update_lora_weights: self.actor_lora_constructor = LoRAConstructor() self.critic_lora_constructor = LoRAConstructor() self.target_auto_balance = False self._target_idx = 0 if self._debug: print(f"[maker{get_rank()}] will send items to {self._detached_trainer_name_list}") if not self._is_fully_initialized: print(f"[maker{get_rank()}] Waiting for INIT") def _get_ready(self): while not self._fully_initialized(): time.sleep(1.0) def _fully_initialized(self): return self._is_fully_initialized def _init_target_trainer_list(self): if len(self.target_trainer_list) > 0: return for name in self._detached_trainer_name_list: self.target_trainer_list.append(ray.get_actor(name, namespace=os.environ["RAY_NAMESPACE"])) # copy from ../trainer/base.py @ray.method(concurrency_group="compute") def _make_experience(self, inputs: Union[Tensor, Dict[str, Tensor]]) -> Experience: if isinstance(inputs, Tensor): return self.experience_maker.make_experience(inputs, **self.generate_kwargs) elif isinstance(inputs, dict): return self.experience_maker.make_experience(**inputs, **self.generate_kwargs) else: raise ValueError(f'Unsupported input type "{type(inputs)}"') @ray.method(concurrency_group="experience_io") def _send_items(self, experience: Experience) -> None: self._init_target_trainer_list() items = split_experience_batch(experience) items_per_trainer = [[] for _ in range(len(self.target_trainer_list))] for item in items: items_per_trainer[self._target_idx].append(item) self._target_idx = (self._target_idx + 1) % len(self.target_trainer_list) for i, target_trainer in enumerate(self.target_trainer_list): if len(items_per_trainer[i]) > 0: target_trainer.buffer_extend.remote(items_per_trainer[i]) def _inference_step(self, batch) -> None: self._on_batch_start() with self._model_visit_lock: self._on_make_experience_start() experience = self._make_experience(batch) self._on_make_experience_end(experience) self._on_send_start() if self.buffer_cpu_offload: experience.to_device("cpu") self._send_items(experience) self._on_send_end() self._on_batch_end() def workingloop(self, dataloader_fn: Callable[[], Iterable], num_epochs: int = 1, num_steps: int = 0): """Working loop of the experience maker. Args: dataloader_fn (Callable[[], Iterable]): A function that returns a dataloader. num_epochs (int, optional): Iterate the dataloader for number of epochs. Defaults to 1. num_steps (int, optional): Iterate the dataloader for number if steps. If this value > 0, num_epochs will be ignored. Defaults to 0. """ self._get_ready() self._on_loop_start() dataloader = dataloader_fn() if num_steps > 0: # ignore num epochs it = iter(dataloader) for _ in tqdm(range(num_steps), desc="ExperienceMaker", disable=not is_rank_0()): try: batch = next(it) except StopIteration: it = iter(dataloader) batch = next(it) self._inference_step(batch) else: with tqdm(total=num_epochs * len(dataloader), desc="ExperienceMaker", disable=not is_rank_0()) as pbar: for _ in range(num_epochs): for batch in dataloader: self._inference_step(batch) pbar.update() self._on_loop_end() @ray.method(concurrency_group="model_io") def update_experience_maker( self, new_actor_state_dict: Dict[str, Any] = None, new_actor_lora_config_dict: Dict[str, Any] = None, new_critic_state_dict: Dict[str, Any] = None, new_critic_lora_config_dict: Dict[str, Any] = None, fully_update: bool = False, chunk_start: bool = None, chunk_end: bool = None, ): """ called by trainer chunk_start: Set True at the first call. Before sending state_dict calls chunk_end: Set True at the last call. After sending state_dict calls. fully_update: Set True if you want to sync models when initializing TODO: load_state_dict integrate with model-sharding strategy """ _watch_memory = self._debug if chunk_start: if self._debug: print("[maker] UPDATE ") if _watch_memory: tracemalloc.start() self._model_visit_lock.acquire() with torch.no_grad(): if new_actor_state_dict is not None: if not self._update_lora_weights or fully_update: self.experience_maker.actor.model.load_state_dict(new_actor_state_dict, strict=False) else: new_actor_state_dict = state_dict_to(new_actor_state_dict, device=torch.cuda.current_device()) state_dict_increase = self.actor_lora_constructor.reconstruct_increase( new_actor_state_dict, new_actor_lora_config_dict ) self.actor_lora_constructor.load_state_dict_increase( self.experience_maker.actor.model, state_dict_increase ) if new_critic_state_dict is not None: if not self._update_lora_weights or fully_update: self.experience_maker.critic.load_state_dict(new_critic_state_dict, strict=False) else: new_critic_state_dict = state_dict_to(new_critic_state_dict, device=torch.cuda.current_device()) state_dict_increase = self.critic_lora_constructor.reconstruct_increase( new_critic_state_dict, new_critic_lora_config_dict ) self.critic_lora_constructor.load_state_dict_increase( self.experience_maker.critic, state_dict_increase ) # the lock must be released after both actor and critic being updated if chunk_end: self._model_visit_lock.release() if _watch_memory: current, peak = tracemalloc.get_traced_memory() print(f"Current memory usage is {current / 10**6}MB; Peak was {peak / 10**6}MB") tracemalloc.stop() if fully_update: self._is_fully_initialized = True def _on_make_experience_start(self) -> None: for callback in self.callbacks: callback.on_make_experience_start() def _on_make_experience_end(self, experience: Experience) -> None: for callback in self.callbacks: callback.on_make_experience_end(experience) def _on_loop_start(self) -> None: for callback in self.callbacks: callback.on_loop_start() def _on_loop_end(self) -> None: for callback in self.callbacks: callback.on_loop_end() def _on_send_start(self) -> None: for callback in self.callbacks: callback.on_send_start() def _on_send_end(self) -> None: for callback in self.callbacks: callback.on_send_end() def _on_batch_start(self) -> None: for callback in self.callbacks: callback.on_batch_start() def _on_batch_end(self) -> None: for callback in self.callbacks: callback.on_batch_end() def _set_default_generate_kwargs(generate_kwargs: dict, actor: Actor) -> None: origin_model = actor.model new_kwargs = {**generate_kwargs} # use huggingface models method directly if "prepare_inputs_fn" not in generate_kwargs and hasattr(origin_model, "prepare_inputs_for_generation"): new_kwargs["prepare_inputs_fn"] = origin_model.prepare_inputs_for_generation if "update_model_kwargs_fn" not in generate_kwargs and hasattr(origin_model, "_update_model_kwargs_for_generation"): new_kwargs["update_model_kwargs_fn"] = origin_model._update_model_kwargs_for_generation return new_kwargs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/__init__.py
applications/ColossalChat/coati/ray/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/lora_constructor.py
applications/ColossalChat/coati/ray/lora_constructor.py
from collections import OrderedDict from dataclasses import dataclass from typing import Any, Dict import torch.nn as nn from coati.models.lora import LoraLinear @dataclass class LoRAConfig: r: int = 0 lora_alpha: int = 1 lora_dropout: float = 0 fan_in_fan_out: bool = False class LoRAConstructor: """ Tools for reconstructing a model from a remote LoRA model. (Transferring only LoRA data costs much less!) Usage: Step 1 (Sender): filter_state_dict_lora() Step 2 (Sender, Optional): extract_lora_config() Step 3 (Sender): send state_dict_lora and lora_config_dict Step 4 (Receiver): reconstruct_increase() Step 5 (Receiver): load_state_dict_increase() """ def __init__(self): self.lora_config_dict = None def register_lora_config(self, lora_config_dict: Dict[str, Any]): self.lora_config_dict = lora_config_dict def reconstruct_increase(self, state_dict_lora: Dict[str, Any], lora_config_dict: Dict[str, Any]): """ xxx.lora_A, xxx.lora_B -->> xxx.weight Warning: the xxx.weight here is the increment actually. """ if lora_config_dict is not None: self.register_lora_config(lora_config_dict) state_dict_increase = OrderedDict() config_iter = iter(self.lora_config_dict.items()) lora_A, lora_B, layer_prefix = None, None, None for k, v in state_dict_lora.items(): if k.rpartition(".")[-1] == "lora_A": lora_A = v layer_prefix = k.rpartition(".")[0] elif k.rpartition(".")[-1] == "lora_B": assert layer_prefix == k.rpartition(".")[0], "unmatched (lora_A, lora_B) pair" layer_prefix_2, config = next(config_iter) assert layer_prefix_2 == layer_prefix, "unmatched (state_dict, config_dict) pair" lora_B = v weight_data_increase = self._compute(lora_A, lora_B, config) state_dict_increase[layer_prefix + ".weight"] = weight_data_increase lora_A, lora_B, layer_prefix = None, None, None else: raise ValueError("unexpected key") return state_dict_increase def _compute(self, lora_A, lora_B, config=LoRAConfig()): def T(w): return w.T if config.fan_in_fan_out else w if config.r > 0: scaling = config.lora_alpha / config.r weight_data_increase = T(lora_B @ lora_A) * scaling return weight_data_increase return 0 def load_state_dict_increase(self, model: nn.Module, state_dict_increase: Dict[str, Any]): """ The final reconstruction step """ # naive approach model.load_state_dict({k: v + model.state_dict()[k] for k, v in state_dict_increase.items()}, strict=False) @staticmethod def filter_state_dict_lora(state_dict: Dict[str, Any], keep_non_lora=False): """ if keep_non_lora, also return non_lora state_dict """ state_dict_lora = OrderedDict() state_dict_non_lora = OrderedDict() for k, v in state_dict.items(): if "lora_A" in k or "lora_B" in k: state_dict_lora[k] = v elif keep_non_lora: state_dict_non_lora[k] = v if keep_non_lora: return state_dict_lora, state_dict_non_lora else: return state_dict_lora, None @staticmethod def extract_lora_config(model: nn.Module) -> Dict[str, LoRAConfig]: """ extract LoraLinear model. return OrderedDict(): name -> LoRAConfig """ lora_config_dict = OrderedDict() for name, child in model.named_modules(): if isinstance(child, LoraLinear): lora_config_dict[name] = LoRAConfig( r=child.r, lora_alpha=child.lora_alpha, lora_dropout=child.lora_dropout, fan_in_fan_out=child.fan_in_fan_out, ) return lora_config_dict
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/detached_replay_buffer.py
applications/ColossalChat/coati/ray/detached_replay_buffer.py
from typing import List import torch from coati.experience_buffer.utils import BufferItem, make_experience_batch, split_experience_batch from coati.experience_maker.base import Experience # from torch.multiprocessing import Queue from ray.util.queue import Queue class DetachedReplayBuffer: """ Detached replay buffer. Share Experience across workers on the same node. Therefore, a trainer node is expected to have only one instance. It is ExperienceMakerHolder's duty to call append(exp) method, remotely. Args: sample_batch_size: Batch size when sampling. Exp won't enqueue until they formed a batch. tp_world_size: Number of workers in the same tp group limit: Limit of number of experience sample BATCHs. A number <= 0 means unlimited. Defaults to 0. cpu_offload: Whether to offload experience to cpu when sampling. Defaults to True. """ def __init__(self, sample_batch_size: int, limit: int = 0) -> None: self.sample_batch_size = sample_batch_size self.limit = limit self.items = Queue(self.limit, actor_options={"num_cpus": 1}) self.batch_collector: List[BufferItem] = [] @torch.no_grad() def append(self, experience: Experience) -> None: """ Expected to be called remotely. """ items = split_experience_batch(experience) self.extend(items) @torch.no_grad() def extend(self, items: List[BufferItem]) -> None: """ Expected to be called remotely. """ self.batch_collector.extend(items) while len(self.batch_collector) >= self.sample_batch_size: items = self.batch_collector[: self.sample_batch_size] experience = make_experience_batch(items) self.items.put(experience, block=True) self.batch_collector = self.batch_collector[self.sample_batch_size :] def clear(self) -> None: # self.items.close() self.items.shutdown() self.items = Queue(self.limit) self.worker_state = [False] * self.tp_world_size self.batch_collector = [] @torch.no_grad() def sample(self, worker_rank=0, to_device="cpu") -> Experience: ret = self._sample_and_erase() ret.to_device(to_device) return ret @torch.no_grad() def _sample_and_erase(self) -> Experience: ret = self.items.get(block=True) return ret def get_length(self) -> int: ret = self.items.qsize() return ret
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/callbacks/__init__.py
applications/ColossalChat/coati/ray/callbacks/__init__.py
from .base import MakerCallback, TrainerCallback from .performance_evaluator import ExperienceMakerPerformanceEvaluator, TrainerPerformanceEvaluator __all__ = [ "TrainerCallback", "MakerCallback", "ExperienceMakerPerformanceEvaluator", "TrainerPerformanceEvaluator", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/callbacks/base.py
applications/ColossalChat/coati/ray/callbacks/base.py
from abc import ABC from coati.experience_maker import Experience class TrainerCallback(ABC): """ Base callback class. It defines the interface for callbacks. """ def on_fit_start(self) -> None: pass def on_fit_end(self) -> None: pass def on_episode_start(self, episode: int) -> None: pass def on_episode_end(self, episode: int) -> None: pass def on_epoch_start(self, epoch: int) -> None: pass def on_epoch_end(self, epoch: int) -> None: pass def on_batch_start(self) -> None: pass def on_batch_end(self, metrics: dict, experience: Experience) -> None: pass def on_update_start(self) -> None: pass def on_update_end(self) -> None: pass class MakerCallback(ABC): def on_loop_start(self) -> None: pass def on_loop_end(self) -> None: pass def on_make_experience_start(self) -> None: pass def on_make_experience_end(self, experience: Experience) -> None: pass def on_send_start(self) -> None: pass def on_send_end(self) -> None: pass def on_batch_start(self) -> None: pass def on_batch_end(self) -> None: pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/coati/ray/callbacks/performance_evaluator.py
applications/ColossalChat/coati/ray/callbacks/performance_evaluator.py
from time import time from typing import Optional import torch import torch.distributed as dist from coati.experience_maker import Experience from .base import MakerCallback, TrainerCallback def get_world_size() -> int: if dist.is_initialized(): return dist.get_world_size() return 1 def print_rank_0(*args, **kwargs) -> None: if not dist.is_initialized() or dist.get_rank() == 0: print(*args, **kwargs) @torch.no_grad() def all_reduce_mean(x: float, world_size: int) -> float: if world_size == 1: return x tensor = torch.tensor([x], device=torch.cuda.current_device()) dist.all_reduce(tensor) tensor = tensor / world_size return tensor.item() class Timer: def __init__(self) -> None: self.start_time: Optional[float] = None self.duration: float = 0.0 def start(self) -> None: self.start_time = time() def end(self) -> None: self.duration += time() - self.start_time def reset(self) -> None: self.duration = 0.0 class ExperienceMakerPerformanceEvaluator(MakerCallback): def __init__( self, actor_num_params: int, critic_num_params: int, initial_model_num_params: int, reward_model_num_params: int ) -> None: super().__init__() self.world_size = get_world_size() self.actor_num_params = actor_num_params self.critic_num_params = critic_num_params self.initial_model_num_params = initial_model_num_params self.reward_model_num_params = reward_model_num_params self.batch_timer = Timer() self.send_timer = Timer() self.make_experience_timer = Timer() self.total_samples: int = 0 self.make_experience_flop: int = 0 print_rank_0( f"ExperienceMaker actor: {actor_num_params/1024**3:.2f}B, critic: {critic_num_params/1024**3:.2f}B, initial model: {initial_model_num_params/1024**3:.2f}B, reward model: {reward_model_num_params/1024**3:.2f}B, world size: {self.world_size}" ) def on_make_experience_start(self) -> None: self.make_experience_timer.start() def on_make_experience_end(self, experience: Experience) -> None: self.make_experience_timer.end() batch_size, seq_len = experience.sequences.shape self.total_samples += batch_size # actor generate num_actions = experience.action_mask.size(1) input_len = seq_len - num_actions total_seq_len = (input_len + seq_len - 1) * num_actions / 2 self.make_experience_flop += self.actor_num_params * batch_size * total_seq_len * 2 # actor forward self.make_experience_flop += self.actor_num_params * batch_size * seq_len * 2 # critic forward self.make_experience_flop += self.critic_num_params * batch_size * seq_len * 2 # initial model forward self.make_experience_flop += self.initial_model_num_params * batch_size * seq_len * 2 # reward model forward self.make_experience_flop += self.reward_model_num_params * batch_size * seq_len * 2 def on_send_start(self) -> None: self.send_timer.start() def on_send_end(self) -> None: self.send_timer.end() def on_batch_start(self) -> None: self.batch_timer.start() def on_batch_end(self) -> None: self.batch_timer.end() def on_loop_end(self) -> None: avg_make_experience_duration = all_reduce_mean(self.make_experience_timer.duration, self.world_size) avg_overall_duration = all_reduce_mean(self.batch_timer.duration, self.world_size) avg_send_duration = all_reduce_mean(self.send_timer.duration, self.world_size) avg_throughput = self.total_samples * self.world_size / (avg_overall_duration + 1e-12) avg_make_experience_tflops = self.make_experience_flop / 1e12 / (avg_make_experience_duration + 1e-12) avg_time_per_sample = (avg_overall_duration + 1e-12) / (self.total_samples * self.world_size) avg_make_experience_time_per_sample = (avg_make_experience_duration + 1e-12) / ( self.total_samples * self.world_size ) avg_send_time_per_sample = (avg_send_duration + 1e-12) / (self.total_samples * self.world_size) print_rank_0( "Making Experience Performance Summary:\n" + f"Throughput: {avg_throughput:.3f} samples/sec\n" + f"TFLOPS per GPU: {avg_make_experience_tflops:.3f}\n" + f"Sample time (overall): {avg_time_per_sample:.3f} s\n" + f"Sample time (make experience): {avg_make_experience_time_per_sample:.3f} s, {avg_make_experience_time_per_sample/avg_time_per_sample*100:.2f}%\n" + f"Sample time (send): {avg_send_time_per_sample:.3f} s, {avg_send_time_per_sample/avg_time_per_sample*100:.2f}%\n" ) class TrainerPerformanceEvaluator(TrainerCallback): def __init__( self, actor_num_params: int, critic_num_params: int, enable_grad_checkpoint: bool = False, ignore_first_episodes: int = 1, ) -> None: super().__init__() self.world_size = get_world_size() self.actor_num_params = actor_num_params self.critic_num_params = critic_num_params self.enable_grad_checkpoint = enable_grad_checkpoint self.ignore_first_episodes = ignore_first_episodes self.ignore_this_episode = False self.episode_timer = Timer() self.batch_timer = Timer() self.update_timer = Timer() self.total_samples: int = 0 self.learn_flop: int = 0 print_rank_0( f"Trainer actor: {self.actor_num_params/1024**3:.2f}B, critic: {self.critic_num_params/1024**3:.2f}B, world size: {self.world_size}" ) def on_episode_start(self, episodes: int) -> None: self.ignore_this_episode = episodes < self.ignore_first_episodes if self.ignore_this_episode: return self.episode_timer.start() def on_episode_end(self, episodes: int) -> None: if self.ignore_this_episode: return self.episode_timer.end() def on_batch_start(self) -> None: if self.ignore_this_episode: return self.batch_timer.start() def on_batch_end(self, metrics: dict, experience: Experience) -> None: if self.ignore_this_episode: return self.batch_timer.end() batch_size, seq_len = experience.sequences.shape self.total_samples += batch_size # actor forward-backward, 3 means forward(1) + backward(2) self.learn_flop += self.actor_num_params * batch_size * seq_len * 2 * (3 + int(self.enable_grad_checkpoint)) # critic forward-backward self.learn_flop += self.critic_num_params * batch_size * seq_len * 2 * (3 + int(self.enable_grad_checkpoint)) def on_update_start(self) -> None: if self.ignore_this_episode: return self.update_timer.start() def on_update_end(self) -> None: if self.ignore_this_episode: return self.update_timer.end() def on_fit_end(self) -> None: if self.total_samples == 0: print_rank_0("No samples are collected, skip trainer performance evaluation") return avg_train_duration = all_reduce_mean(self.batch_timer.duration, self.world_size) avg_update_duration = all_reduce_mean(self.update_timer.duration, self.world_size) avg_episode_duration = all_reduce_mean(self.episode_timer.duration, self.world_size) avg_throughput = self.total_samples * self.world_size / (avg_episode_duration + 1e-12) avg_learn_tflops = self.learn_flop / 1e12 / (avg_train_duration + 1e-12) avg_time_per_sample = (avg_episode_duration + 1e-12) / (self.total_samples * self.world_size) avg_train_time_per_sample = (avg_train_duration + 1e-12) / (self.total_samples * self.world_size) avg_update_time_per_sample = (avg_update_duration + 1e-12) / (self.total_samples * self.world_size) print_rank_0( "Learning Performance Summary:\n" + f"Throughput: {avg_throughput:.3f} samples/sec\n" + f"TFLOPS per GPU: {avg_learn_tflops:.3f}\n" + f"Sample time (overall): {avg_time_per_sample:.3f} s\n" + f"Sample time (train): {avg_train_time_per_sample:.3f} s, {avg_train_time_per_sample/avg_time_per_sample*100:.2f}%\n" + f"Sample time (update): {avg_update_time_per_sample:.3f} s, {avg_update_time_per_sample/avg_time_per_sample*100:.2f}%\n" )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/benchmarks/dummy_dataset.py
applications/ColossalChat/benchmarks/dummy_dataset.py
from typing import Callable from torch.utils.data import Dataset class DummyLLMDataset(Dataset): def __init__(self, keys, seq_len, size=500, gen_fn={}): self.keys = keys self.gen_fn = gen_fn self.seq_len = seq_len self.data = self._generate_data() self.size = size def _generate_data(self): data = {} for key in self.keys: if key in self.gen_fn: data[key] = self.gen_fn[key] else: data[key] = [1] * self.seq_len return data def __len__(self): return self.size def __getitem__(self, idx): return { key: self.data[key] if not isinstance(self.data[key], Callable) else self.data[key](idx) for key in self.keys }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/benchmarks/prepare_dummy_test_dataset.py
applications/ColossalChat/benchmarks/prepare_dummy_test_dataset.py
import argparse import json import os import time from multiprocessing import cpu_count from datasets import load_dataset from dummy_dataset import DummyLLMDataset from colossalai.logging import get_dist_logger logger = get_dist_logger() if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--data_dir", type=str, required=True, default=None, help="The output dir", ) parser.add_argument( "--dataset_size", type=int, required=True, default=None, help="The size of data", ) parser.add_argument( "--max_length", type=int, required=True, default=None, help="The max length of data", ) parser.add_argument( "--data_type", type=str, required=True, default=None, help="The type of data, choose one from ['sft', 'prompt', 'preference', 'kto']", ) args = parser.parse_args() if args.data_type == "sft": dataset = DummyLLMDataset(["input_ids", "attention_mask", "labels"], args.max_length, args.dataset_size) elif args.data_type == "prompt": # pass PPO dataset is prepared separately pass elif args.data_type == "preference": dataset = DummyLLMDataset( ["chosen_input_ids", "chosen_loss_mask", "rejected_input_ids", "rejected_loss_mask"], args.max_length, args.dataset_size, ) elif args.data_type == "kto": dataset = DummyLLMDataset( ["prompt", "completion", "label"], args.max_length - 512, args.dataset_size, gen_fn={ "completion": lambda x: [1] * 512, "label": lambda x: x % 2, }, ) else: raise ValueError(f"Unknown data type {args.data_type}") # Save each jsonl spliced dataset. output_index = "0" output_name = f"part-{output_index}" os.makedirs(args.data_dir, exist_ok=True) output_jsonl_path = os.path.join(args.data_dir, "json") output_arrow_path = os.path.join(args.data_dir, "arrow") output_cache_path = os.path.join(args.data_dir, "cache") os.makedirs(output_jsonl_path, exist_ok=True) os.makedirs(output_arrow_path, exist_ok=True) output_jsonl_file_path = os.path.join(output_jsonl_path, output_name + ".jsonl") st = time.time() with open(file=output_jsonl_file_path, mode="w", encoding="utf-8") as fp_writer: count = 0 for i in range(len(dataset)): data_point = dataset[i] if count % 500 == 0: logger.info(f"processing {count} spliced data points for {fp_writer.name}") count += 1 fp_writer.write(json.dumps(data_point, ensure_ascii=False) + "\n") logger.info( f"Current file {fp_writer.name}; " f"Data size: {len(dataset)}; " f"Time cost: {round((time.time() - st) / 60, 6)} minutes." ) # Save each arrow spliced dataset output_arrow_file_path = os.path.join(output_arrow_path, output_name) logger.info(f"Start to save {output_arrow_file_path}") dataset = load_dataset( path="json", data_files=[output_jsonl_file_path], cache_dir=os.path.join(output_cache_path, "tokenized"), keep_in_memory=False, num_proc=cpu_count(), split="train", ) dataset.save_to_disk(dataset_path=output_arrow_file_path, num_proc=min(len(dataset), cpu_count()))
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/benchmarks/benchmark_ppo.py
applications/ColossalChat/benchmarks/benchmark_ppo.py
""" For becnhmarking ppo. Mudified from examples/training_scripts/train_ppo.py """ import argparse import json import os import resource from contextlib import nullcontext import torch import torch.distributed as dist from coati.dataset import ( DataCollatorForPromptDataset, DataCollatorForSupervisedDataset, StatefulDistributedSampler, load_tokenized_dataset, setup_conversation_template, setup_distributed_dataloader, ) from coati.models import Critic, RewardModel, convert_to_lora_module, disable_dropout from coati.trainer import PPOTrainer from coati.trainer.callbacks import PerformanceEvaluator from coati.trainer.utils import is_rank_0 from coati.utils import load_checkpoint, replace_with_flash_attention from transformers import AutoTokenizer, OPTForCausalLM from transformers.models.opt.configuration_opt import OPTConfig import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin from colossalai.cluster import DistCoordinator from colossalai.lazy import LazyInitContext from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam from colossalai.utils import get_current_device def get_model_numel(model: torch.nn.Module, plugin: str, tp: int) -> int: numel = sum(p.numel() for p in model.parameters()) if plugin == "3d" and tp > 1: numel *= dist.get_world_size() return numel def get_gpt_config(model_name: str) -> OPTConfig: model_map = { "125m": OPTConfig.from_pretrained("facebook/opt-125m"), "350m": OPTConfig(hidden_size=1024, ffn_dim=4096, num_hidden_layers=24, num_attention_heads=16), "700m": OPTConfig(hidden_size=1280, ffn_dim=5120, num_hidden_layers=36, num_attention_heads=20), "1.3b": OPTConfig.from_pretrained("facebook/opt-1.3b"), "2.7b": OPTConfig.from_pretrained("facebook/opt-2.7b"), "3.5b": OPTConfig(hidden_size=3072, ffn_dim=12288, num_hidden_layers=32, num_attention_heads=32), "5.5b": OPTConfig(hidden_size=3840, ffn_dim=15360, num_hidden_layers=32, num_attention_heads=32), "6.7b": OPTConfig.from_pretrained("facebook/opt-6.7b"), "10b": OPTConfig(hidden_size=5120, ffn_dim=20480, num_hidden_layers=32, num_attention_heads=32), "13b": OPTConfig.from_pretrained("facebook/opt-13b"), } try: return model_map[model_name] except KeyError: raise ValueError(f'Unknown model "{model_name}"') def benchmark_train(args): # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== init_ctx = LazyInitContext(default_device=get_current_device()) if "gemini" in args.plugin else nullcontext() booster_policy = None with init_ctx: actor = OPTForCausalLM(config=get_gpt_config(args.pretrain)) # Disable dropout disable_dropout(actor) ref_model = OPTForCausalLM(config=get_gpt_config(args.pretrain)) reward_model = RewardModel(config=get_gpt_config("350m")) critic = Critic(config=get_gpt_config("350m")) disable_dropout(critic) actor_numel = get_model_numel(actor, args.plugin, args.tp) critic_numel = get_model_numel(critic, args.plugin, args.tp) initial_model_numel = get_model_numel(ref_model, args.plugin, args.tp) reward_model_numel = get_model_numel(reward_model, args.plugin, args.tp) performance_evaluator = PerformanceEvaluator( actor_numel, critic_numel, initial_model_numel, reward_model_numel, enable_grad_checkpoint=False, ignore_episodes=2, train_config={"model": "facebook/opt-" + args.pretrain, "lora_rank": args.lora_rank, "plugin": args.plugin}, save_path="./benchmark_performance_summarization.txt", ) if args.tp > 1: if reward_model.model.config.architectures[0] != critic.model.config.architectures[0]: raise ValueError("Reward model and critic model must have the same architecture") if reward_model.model.config.architectures[0] == "BloomForCausalLM": from colossalai.shardformer.policies.bloom import BloomPolicy booster_policy = BloomPolicy() elif reward_model.model.config.architectures[0] == "LlamaForCausalLM": from colossalai.shardformer.policies.llama import LlamaPolicy booster_policy = LlamaPolicy() elif reward_model.model.config.architectures[0] == "GPT2LMHeadModel": from colossalai.shardformer.policies.gpt2 import GPT2Policy booster_policy = GPT2Policy() elif reward_model.model.config.architectures[0] == "ChatGLMModel": from colossalai.shardformer.policies.chatglm2 import ChatGLMPolicy booster_policy = ChatGLMPolicy() elif reward_model.model.config.architectures[0] == "OPTForCausalLM": from colossalai.shardformer.policies.opt import OPTPolicy booster_policy = OPTPolicy() else: raise ValueError("Unknown model architecture for policy") if args.lora_rank > 0: actor = convert_to_lora_module(actor, args.lora_rank, lora_train_bias=args.lora_train_bias) critic = convert_to_lora_module(critic, args.lora_rank, lora_train_bias=args.lora_train_bias) if args.grad_checkpoint and args.lora_rank == 0: actor.gradient_checkpointing_enable() critic.model.gradient_checkpointing_enable() coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") elif args.lora_rank > 0: coordinator.print_on_master(msg="Gradient checkpointing will be disabled when LoRA is enabled") if args.use_flash_attn: replace_with_flash_attention(model=actor) replace_with_flash_attention(model=critic) coordinator.print_on_master(msg="Flash-attention enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir) if os.path.exists(args.conversation_template_config): conversation_template_config = json.load(open(args.conversation_template_config, "r", encoding="utf8")) conversation_template = setup_conversation_template( tokenizer, chat_template_config=conversation_template_config, save_path=args.conversation_template_config ) stop_token_ids = ( conversation_template.assistant_line_end if len(conversation_template.assistant_line_end) > 0 else None ) else: raise ValueError("Conversation template config is not provided or incorrect") if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False tokenizer.padding_side = "left" # left padding for generation (online learning) # configure generation config actor.generation_config.update( pad_token_id=tokenizer.eos_token_id, bos_token_id=tokenizer.bos_token_id, eos_token_id=tokenizer.eos_token_id ) # configure optimizer coordinator.print_on_master(f"setting up optimizer for actor: lr={args.lr}, weight_decay={args.weight_decay}") actor_optim = HybridAdam( model_params=actor.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) coordinator.print_on_master(f"setting up optimizer for critic: lr={args.lr}, weight_decay={args.weight_decay}") critic_optim = HybridAdam( model_params=critic.parameters(), lr=args.critic_lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # configure dataset coordinator.print_on_master(f"Load dataset: {args.prompt_dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_prompt_dataset = load_tokenized_dataset(dataset_paths=args.prompt_dataset, mode="train", mode_map=mode_map) coordinator.print_on_master(f"prompt dataset size: {len(train_prompt_dataset)}") data_collator = DataCollatorForPromptDataset(tokenizer=tokenizer, max_length=args.max_length - args.max_seq_len) train_prompt_dataloader = setup_distributed_dataloader( dataset=train_prompt_dataset, batch_size=args.experience_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, use_tp=args.tp > 1, ) if len(args.pretrain_dataset) > 0: train_pretrain_dataset = load_tokenized_dataset( dataset_paths=args.pretrain_dataset, mode="train", mode_map=mode_map ) data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer, max_length=args.max_length) train_pretrain_dataloader = setup_distributed_dataloader( dataset=train_pretrain_dataset, batch_size=args.ptx_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, use_tp=args.tp > 1, ) else: train_pretrain_dataloader = None if args.warmup_steps is None: args.warmup_steps = int(0.025 * args.num_episodes) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") actor_lr_scheduler = CosineAnnealingWarmupLR( optimizer=actor_optim, total_steps=args.num_episodes, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) critic_lr_scheduler = CosineAnnealingWarmupLR( optimizer=critic_optim, total_steps=args.num_episodes, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) # ============================== # Initialize Booster # ============================== if args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=1, zero_stage=0, precision=args.mixed_precision, ) custom_plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=1, zero_stage=0, precision=args.mixed_precision, custom_policy=booster_policy, ) else: raise ValueError(f"Unknown plugin {args.plugin}") if args.plugin != "3d": custom_plugin = plugin actor_booster = Booster(plugin=plugin) ref_booster = Booster(plugin=plugin) rm_booster = Booster(plugin=custom_plugin) critic_booster = Booster(plugin=custom_plugin) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) actor, actor_optim, _, train_prompt_dataloader, actor_lr_scheduler = actor_booster.boost( model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, dataloader=train_prompt_dataloader, ) critic, critic_optim, _, _, critic_lr_scheduler = critic_booster.boost( model=critic, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, dataloader=train_prompt_dataloader, ) reward_model, _, _, _, _ = rm_booster.boost(model=reward_model, dataloader=train_prompt_dataloader) ref_model, _, _, _, _ = ref_booster.boost(model=ref_model, dataloader=train_prompt_dataloader) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) sampler_start_idx = 0 start_step = 0 if args.rm_checkpoint_path is not None: if "modeling" in args.rm_checkpoint_path: rm_booster.load_model(reward_model, args.rm_checkpoint_path) else: _, _, _ = load_checkpoint( load_dir=args.rm_checkpoint_path, booster=rm_booster, model=reward_model, optimizer=None, lr_scheduler=None, ) coordinator.print_on_master(f"Loaded reward model checkpoint {args.rm_checkpoint_path}") if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: actor_booster.load_model(actor, args.checkpoint_path) ref_booster.load_model(ref_model, args.checkpoint_path) coordinator.print_on_master(f"Loaded actor and reference model {args.checkpoint_path}") else: _, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=actor_booster, model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, ) _, _, _ = load_checkpoint( load_dir=args.checkpoint_path, booster=ref_booster, model=ref_model, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, ) assert isinstance(train_prompt_dataloader.sampler, StatefulDistributedSampler) train_prompt_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded actor and reference model checkpoint {args.checkpoint_path} at spisode {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) if args.critic_checkpoint_path is not None: if "modeling" in args.critic_checkpoint_path: critic_booster.load_model(critic, args.critic_checkpoint_path) else: _, _, _ = load_checkpoint( load_dir=args.critic_checkpoint_path, booster=critic_booster, model=critic, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, ) coordinator.print_on_master(f"Loaded critic checkpoint {args.critic_checkpoint_path}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) # configure trainer trainer = PPOTrainer( actor_booster, critic_booster, actor, critic, reward_model, ref_model, actor_optim, critic_optim, actor_lr_scheduler, critic_lr_scheduler, tokenizer=tokenizer, stop_token_ids=stop_token_ids, kl_coef=args.kl_coef, ptx_coef=args.ptx_coef, train_batch_size=args.train_batch_size, buffer_limit=args.num_collect_steps * args.experience_batch_size, max_length=args.max_length, max_new_tokens=args.max_seq_len, use_cache=True, do_sample=True, temperature=0.7, accumulation_steps=args.accumulation_steps, save_dir=args.save_path, save_interval=args.save_interval, top_k=50, use_tp=args.tp > 1, offload_inference_models="gemini" not in args.plugin, callbacks=[performance_evaluator], coordinator=coordinator, ) trainer.fit( num_episodes=args.num_episodes, num_collect_steps=args.num_collect_steps, num_update_steps=args.num_update_steps, prompt_dataloader=train_prompt_dataloader, pretrain_dataloader=train_pretrain_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if args.lora_rank > 0 and args.merge_lora_weights: from coati.models.lora import LORA_MANAGER # NOTE: set model to eval to merge LoRA weights LORA_MANAGER.merge_weights = True actor.eval() critic.eval() # save model checkpoint after fitting on only rank0 coordinator.print_on_master("Start saving final actor model checkpoint") actor_booster.save_model(actor, os.path.join(trainer.actor_save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final actor model checkpoint at episodes {args.num_episodes} at folder {args.save_path}" ) coordinator.print_on_master("Start saving final critic model checkpoint") critic_booster.save_model(critic, os.path.join(trainer.critic_save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final critic model checkpoint at episodes {args.num_episodes} at folder {args.save_path}" ) memory_consumption = torch.cuda.max_memory_allocated() / 1024**2 if is_rank_0(): with open("./benchmark_memory_consumption.txt", "a+") as f: f.write( f"Model=Opt-{args.pretrain}; lora_rank={args.lora_rank}; plugin={args.plugin}\nMax CUDA memory usage: {memory_consumption:.2f} MB\n" ) coordinator.print_on_master(f"Max CUDA memory usage: {memory_consumption:.2f} MB") if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_dataset", nargs="+", default=[]) parser.add_argument("--pretrain_dataset", nargs="+", default=[]) parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d"], help="Choose which plugin to use", ) parser.add_argument( "--conversation_template_config", type=str, default=None, help="Path \ to save conversation template config files.", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--checkpoint_path", type=str, default=None) parser.add_argument("--critic_checkpoint_path", type=str, default=None) parser.add_argument("--rm_checkpoint_path", type=str, help="Reward model checkpoint path") parser.add_argument("--save_path", type=str, default="actor_checkpoint_prompts") parser.add_argument("--num_episodes", type=int, default=1) parser.add_argument("--num_collect_steps", type=int, default=2) parser.add_argument("--num_update_steps", type=int, default=5) parser.add_argument("--save_interval", type=int, default=1000) parser.add_argument("--train_batch_size", type=int, default=16) parser.add_argument("--experience_batch_size", type=int, default=16) parser.add_argument("--ptx_batch_size", type=int, default=1) parser.add_argument("--lora_train_bias", type=str, default="none") parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--lora_rank", type=int, default=0, help="low-rank adaptation matrices rank") parser.add_argument("--merge_lora_weights", type=bool, default=True) parser.add_argument("--lr", type=float, default=9e-6) parser.add_argument("--critic_lr", type=float, default=9e-6) parser.add_argument("--kl_coef", type=float, default=0.1) parser.add_argument("--ptx_coef", type=float, default=0.0) parser.add_argument("--max_length", type=int, default=512) parser.add_argument("--max_seq_len", type=int, default=256) parser.add_argument("--log_dir", default="logs", type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() benchmark_train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/benchmarks/ray/1mmt_dummy.py
applications/ColossalChat/benchmarks/ray/1mmt_dummy.py
import argparse import os import socket from functools import partial import ray import torch from coati.quant import llama_load_quant, low_resource_init from coati.ray.detached_trainer_ppo import DetachedPPOTrainer from coati.ray.experience_maker_holder import ExperienceMakerHolder from coati.ray.utils import ( get_actor_from_args, get_critic_from_args, get_receivers_per_sender, get_reward_model_from_args, get_strategy_from_args, ) from torch.utils.data import DataLoader from transformers import AutoConfig, AutoTokenizer from transformers.modeling_utils import no_init_weights def get_free_port(): with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: s.bind(("", 0)) return s.getsockname()[1] def get_local_ip(): with socket.socket(socket.AF_INET, socket.SOCK_DGRAM) as s: s.connect(("8.8.8.8", 80)) return s.getsockname()[0] def main(args): master_addr = str(get_local_ip()) # trainer_env_info trainer_port = str(get_free_port()) env_info_trainers = [ { "local_rank": "0", "rank": str(rank), "world_size": str(args.num_trainers), "master_port": trainer_port, "master_addr": master_addr, } for rank in range(args.num_trainers) ] # maker_env_info maker_port = str(get_free_port()) env_info_maker = { "local_rank": "0", "rank": "0", "world_size": "1", "master_port": maker_port, "master_addr": master_addr, } # configure tokenizer tokenizer = AutoTokenizer.from_pretrained(args.pretrain) tokenizer.pad_token = tokenizer.eos_token def model_fn(): actor_cfg = AutoConfig.from_pretrained(args.pretrain) critic_cfg = AutoConfig.from_pretrained(args.critic_pretrain) actor = get_actor_from_args(args.model, config=actor_cfg).requires_grad_(False).half().cuda() critic = get_critic_from_args(args.critic_model, config=critic_cfg).requires_grad_(False).half().cuda() reward_model = ( get_reward_model_from_args(args.critic_model, config=critic_cfg).requires_grad_(False).half().cuda() ) if args.initial_model_quant_ckpt is not None and args.model == "llama": # quantize initial model with low_resource_init(), no_init_weights(): initial_model = get_actor_from_args(args.model, config=actor_cfg) initial_model.model = ( llama_load_quant( initial_model.model, args.initial_model_quant_ckpt, args.quant_bits, args.quant_group_size ) .cuda() .requires_grad_(False) ) else: initial_model = get_actor_from_args(args.model, config=actor_cfg).requires_grad_(False).half().cuda() return actor, critic, reward_model, initial_model # configure Experience Maker experience_holder_ref = ExperienceMakerHolder.options(name="maker0", num_gpus=1, max_concurrency=2).remote( detached_trainer_name_list=[f"trainer{i}" for i in range(args.num_trainers)], strategy_fn=partial(get_strategy_from_args, args.maker_strategy), model_fn=model_fn, env_info=env_info_maker, kl_coef=0.1, debug=args.debug, # sync_models_from_trainers=True, # generation kwargs: max_length=512, do_sample=True, temperature=1.0, top_k=50, pad_token_id=tokenizer.pad_token_id, eos_token_id=tokenizer.eos_token_id, eval_performance=True, use_cache=True, ) def trainer_model_fn(): actor = get_actor_from_args(args.model, config=AutoConfig.from_pretrained(args.pretrain)).half().cuda() critic = ( get_critic_from_args(args.critic_model, config=AutoConfig.from_pretrained(args.critic_pretrain)) .half() .cuda() ) return actor, critic # configure Trainer trainer_refs = [ DetachedPPOTrainer.options(name=f"trainer{i}", num_gpus=1, max_concurrency=2).remote( experience_maker_holder_name_list=[ f"maker{x}" for x in get_receivers_per_sender(i, args.num_trainers, 1, allow_idle_sender=True) ], strategy_fn=partial(get_strategy_from_args, args.trainer_strategy), model_fn=trainer_model_fn, env_info=env_info_trainer, train_batch_size=args.train_batch_size, buffer_limit=16, eval_performance=True, debug=args.debug, ) for i, env_info_trainer in enumerate(env_info_trainers) ] dataset_size = args.experience_batch_size * 4 def data_gen_fn(): input_ids = torch.randint(tokenizer.vocab_size, (256,), device=torch.cuda.current_device()) attn_mask = torch.ones_like(input_ids) return {"input_ids": input_ids, "attention_mask": attn_mask} def build_dataloader(size): dataset = [data_gen_fn() for _ in range(size)] dataloader = DataLoader(dataset, batch_size=args.experience_batch_size) return dataloader # uncomment this function if sync_models_from_trainers is True # ray.get([ # trainer_ref.sync_models_to_remote_makers.remote() # for trainer_ref in trainer_refs # ]) wait_tasks = [] wait_tasks.append( experience_holder_ref.workingloop.remote( partial(build_dataloader, dataset_size), num_steps=args.experience_steps ) ) total_steps = args.experience_batch_size * args.experience_steps // (args.num_trainers * args.train_batch_size) for trainer_ref in trainer_refs: wait_tasks.append(trainer_ref.fit.remote(total_steps, args.update_steps, args.train_epochs)) ray.get(wait_tasks) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--num_trainers", type=int, default=1) parser.add_argument( "--trainer_strategy", choices=["ddp", "colossalai_gemini", "colossalai_zero2", "colossalai_gemini_cpu", "colossalai_zero2_cpu"], default="ddp", ) parser.add_argument("--maker_strategy", choices=["naive"], default="naive") parser.add_argument("--model", default="gpt2", choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--critic_model", default="gpt2", choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--critic_pretrain", type=str, default=None) parser.add_argument("--experience_steps", type=int, default=4) parser.add_argument("--experience_batch_size", type=int, default=8) parser.add_argument("--train_epochs", type=int, default=1) parser.add_argument("--update_steps", type=int, default=2) parser.add_argument("--train_batch_size", type=int, default=8) parser.add_argument("--lora_rank", type=int, default=0, help="low-rank adaptation matrices rank") parser.add_argument("--initial_model_quant_ckpt", type=str, default=None) parser.add_argument("--quant_bits", type=int, default=4) parser.add_argument("--quant_group_size", type=int, default=128) parser.add_argument("--debug", action="store_true") args = parser.parse_args() ray.init(namespace=os.environ["RAY_NAMESPACE"], runtime_env={"env_vars": dict(os.environ)}) main(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/benchmarks/ray/mmmt_dummy.py
applications/ColossalChat/benchmarks/ray/mmmt_dummy.py
import argparse import os import socket from functools import partial import ray import torch from coati.quant import llama_load_quant, low_resource_init from coati.ray.detached_trainer_ppo import DetachedPPOTrainer from coati.ray.experience_maker_holder import ExperienceMakerHolder from coati.ray.utils import ( get_actor_from_args, get_critic_from_args, get_receivers_per_sender, get_reward_model_from_args, get_strategy_from_args, ) from torch.utils.data import DataLoader from transformers import AutoConfig, AutoTokenizer from transformers.modeling_utils import no_init_weights def get_free_port(): with socket.socket(socket.AF_INET, socket.SOCK_STREAM) as s: s.bind(("", 0)) return s.getsockname()[1] def get_local_ip(): with socket.socket(socket.AF_INET, socket.SOCK_DGRAM) as s: s.connect(("8.8.8.8", 80)) return s.getsockname()[0] def main(args): master_addr = str(get_local_ip()) # trainer_env_info trainer_port = str(get_free_port()) env_info_trainers = [ { "local_rank": "0", "rank": str(rank), "world_size": str(args.num_trainers), "master_port": trainer_port, "master_addr": master_addr, } for rank in range(args.num_trainers) ] # maker_env_info maker_port = str(get_free_port()) env_info_makers = [ { "local_rank": "0", "rank": str(rank), "world_size": str(args.num_makers), "master_port": maker_port, "master_addr": master_addr, } for rank in range(args.num_makers) ] # configure tokenizer tokenizer = AutoTokenizer.from_pretrained(args.pretrain) tokenizer.pad_token = tokenizer.eos_token def model_fn(): actor_cfg = AutoConfig.from_pretrained(args.pretrain) critic_cfg = AutoConfig.from_pretrained(args.critic_pretrain) actor = get_actor_from_args(args.model, config=actor_cfg).requires_grad_(False).half().cuda() critic = get_critic_from_args(args.critic_model, config=critic_cfg).requires_grad_(False).half().cuda() reward_model = ( get_reward_model_from_args(args.critic_model, config=critic_cfg).requires_grad_(False).half().cuda() ) if args.initial_model_quant_ckpt is not None and args.model == "llama": # quantize initial model with low_resource_init(), no_init_weights(): initial_model = get_actor_from_args(args.model, config=actor_cfg) initial_model.model = ( llama_load_quant( initial_model.model, args.initial_model_quant_ckpt, args.quant_bits, args.quant_group_size ) .cuda() .requires_grad_(False) ) else: initial_model = get_actor_from_args(args.model, config=actor_cfg).requires_grad_(False).half().cuda() return actor, critic, reward_model, initial_model # configure Experience Maker experience_holder_refs = [ ExperienceMakerHolder.options(name=f"maker{i}", num_gpus=1, max_concurrency=2).remote( detached_trainer_name_list=[ f"trainer{x}" for x in get_receivers_per_sender(i, args.num_makers, args.num_trainers, allow_idle_sender=False) ], strategy_fn=partial(get_strategy_from_args, args.maker_strategy), model_fn=model_fn, env_info=env_info_maker, kl_coef=0.1, debug=args.debug, # sync_models_from_trainers=True, # generation kwargs: max_length=512, do_sample=True, temperature=1.0, top_k=50, pad_token_id=tokenizer.pad_token_id, eos_token_id=tokenizer.eos_token_id, eval_performance=True, use_cache=True, ) for i, env_info_maker in enumerate(env_info_makers) ] def trainer_model_fn(): actor = get_actor_from_args(args.model, config=AutoConfig.from_pretrained(args.pretrain)).half().cuda() critic = ( get_critic_from_args(args.critic_model, config=AutoConfig.from_pretrained(args.critic_pretrain)) .half() .cuda() ) return actor, critic # configure Trainer trainer_refs = [ DetachedPPOTrainer.options(name=f"trainer{i}", num_gpus=1, max_concurrency=2).remote( experience_maker_holder_name_list=[ f"maker{x}" for x in get_receivers_per_sender(i, args.num_trainers, args.num_makers, allow_idle_sender=True) ], strategy_fn=partial(get_strategy_from_args, args.trainer_strategy), model_fn=trainer_model_fn, env_info=env_info_trainer, train_batch_size=args.train_batch_size, buffer_limit=16, eval_performance=True, debug=args.debug, ) for i, env_info_trainer in enumerate(env_info_trainers) ] dataset_size = args.experience_batch_size * 4 def data_gen_fn(): input_ids = torch.randint(tokenizer.vocab_size, (256,), device=torch.cuda.current_device()) attn_mask = torch.ones_like(input_ids) return {"input_ids": input_ids, "attention_mask": attn_mask} def build_dataloader(size): dataset = [data_gen_fn() for _ in range(size)] dataloader = DataLoader(dataset, batch_size=args.experience_batch_size) return dataloader # uncomment this function if sync_models_from_trainers is True # ray.get([ # trainer_ref.sync_models_to_remote_makers.remote() # for trainer_ref in trainer_refs # ]) wait_tasks = [] for experience_holder_ref in experience_holder_refs: wait_tasks.append( experience_holder_ref.workingloop.remote( partial(build_dataloader, dataset_size), num_steps=args.experience_steps ) ) total_steps = ( args.experience_batch_size * args.experience_steps * args.num_makers // (args.num_trainers * args.train_batch_size) ) for trainer_ref in trainer_refs: wait_tasks.append(trainer_ref.fit.remote(total_steps, args.update_steps, args.train_epochs)) ray.get(wait_tasks) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--num_makers", type=int, default=1) parser.add_argument("--num_trainers", type=int, default=1) parser.add_argument( "--trainer_strategy", choices=["ddp", "colossalai_gemini", "colossalai_zero2", "colossalai_gemini_cpu", "colossalai_zero2_cpu"], default="ddp", ) parser.add_argument("--maker_strategy", choices=["naive"], default="naive") parser.add_argument("--model", default="gpt2", choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--critic_model", default="gpt2", choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--critic_pretrain", type=str, default=None) parser.add_argument("--experience_steps", type=int, default=4) parser.add_argument("--experience_batch_size", type=int, default=8) parser.add_argument("--train_epochs", type=int, default=1) parser.add_argument("--update_steps", type=int, default=2) parser.add_argument("--train_batch_size", type=int, default=8) parser.add_argument("--lora_rank", type=int, default=0, help="low-rank adaptation matrices rank") parser.add_argument("--initial_model_quant_ckpt", type=str, default=None) parser.add_argument("--quant_bits", type=int, default=4) parser.add_argument("--quant_group_size", type=int, default=128) parser.add_argument("--debug", action="store_true") args = parser.parse_args() ray.init(namespace=os.environ["RAY_NAMESPACE"], runtime_env={"env_vars": dict(os.environ)}) main(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/tests/verify_chat_data.py
applications/ColossalChat/tests/verify_chat_data.py
import argparse import json if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--data_source", type=str, required=True, default=None, help="The raw data file", ) parser.add_argument( "--to_verify_file", type=str, required=True, default=None, help="The file that contains the data to be verified", ) parser.add_argument( "--data_type", type=str, required=True, default=None, help="The data type", ) args = parser.parse_args() # Read data data = [] with open(args.data_source, "r", encoding="utf8") as f: for line in f.readlines(): data.append(json.loads(line)) to_verify_data = [] with open(args.to_verify_file, "r", encoding="utf8") as f: for line in f.readlines(): to_verify_data.append(json.loads(line)) if args.data_type == "sft": target_lable = [msg["content"].strip() for msg in data[0]["messages"] if msg["from"] == "assistant"] target_negative_label = [msg["content"].strip() for msg in data[0]["messages"] if msg["from"] == "human"] # Read to verify file to_verify_lable = to_verify_data[0]["labels_decode"] for label in target_lable: assert any([label in s for s in to_verify_lable]), f"Label {label} not in target label {to_verify_lable}" for label in target_negative_label: assert all( [label not in s for s in to_verify_lable] ), f"Negative label {label} in target label {to_verify_lable}" elif args.data_type == "dpo": chosen_lable = data[0]["chosen"][0]["content"].strip() rejected_lable = data[0]["rejected"][0]["content"].strip() # Read to verify file to_verify_lable_chosen = to_verify_data[0]["chosen_label_decode"] to_verify_lable_rejected = to_verify_data[0]["rejected_label_decode"] assert any( [chosen_lable in s for s in to_verify_lable_chosen] ), f"Chosen label {chosen_lable} not in target chosen label {to_verify_lable_chosen}" assert any( [rejected_lable in s for s in to_verify_lable_rejected] ), f"Rejected label {rejected_lable} not in target rejected label {to_verify_lable_chosen}" elif args.data_type == "kto": sample = data[0] to_verify_data = to_verify_data[0] for line in sample["prompt"]: assert line["content"] in to_verify_data["input_id_decode"] assert sample["completion"]["content"] in to_verify_data["input_id_decode"] assert sample["completion"]["content"] in to_verify_data["completion_decode"] assert sample["label"] == to_verify_data["label"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/tests/generate_dummy_datasets_for_testing.py
applications/ColossalChat/tests/generate_dummy_datasets_for_testing.py
import argparse import json import os sft_seed = { "messages": [ {"from": "user", "content": "Give three tips for staying healthy."}, { "from": "assistant", "content": "1.Eat a balanced diet and make sure to include plenty of fruits and vegetables. \n2. Exercise regularly to keep your body active and strong. \n3. Get enough sleep and maintain a consistent sleep schedule.", }, ] } prompt_seed = { "messages": [ {"from": "user", "content": "Describe the impacts of climate change on communities living in coastal areas."}, { "from": "assistant", "content": "Climate change has caused an increase in sea levels, which has caused coastal erosion and flooding of low-lying areas. This has led to displacement of people from their homes, as well as increased risk of epidemics of waterborne illnesses. Coastal cities have also seen an increase in extreme weather events such as hurricanes and tropical storms, which can cause extensive damage to infrastructure, homes, and businesses. As a result of climate change, some coastal areas are becoming uninhabitable, forcing communities to seek alternative living arrangements.", }, ] } prompt_rlvr_seed = { "messages": [ { "from": "user", "content": "What is the degree of the polynomial $(4 +5x^3 +100 +2\pi x^4 + \sqrt{10}x^4 +9)$?", }, ], "gt_answer": "4", } preference_seed = { "context": [ {"from": "user", "content": "What kind of noises did dinosaurs make?"}, { "from": "assistant", "content": "Humans and dinosaurs didn't live at the same time, so it's really hard to say. The best place to find out what noises dinosaurs made would be", }, {"from": "user", "content": "yes they did"}, { "from": "assistant", "content": "to guess, and that would probably require lots of reading and a certain amount of imagination, so we're not really prepared to do that.", }, {"from": "user", "content": "you cant read"}, ], "chosen": [{"from": "assistant", "content": "You can read?"}], "rejected": [{"from": "assistant", "content": "there's a lot of stuff humans don't know"}], } kto_seed = { "prompt": [ {"from": "user", "content": "What are some praise words in english?"}, { "from": "assistant", "content": "Here's an incomplete list.\n\nexcellent, fantastic, impressive ...", }, {"from": "user", "content": "What's your favorite one?"}, ], "completion": {"from": "assistant", "content": "Impressive."}, "label": True, } if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "--data_dir", type=str, required=True, default=None, help="The output dir", ) parser.add_argument( "--data_type", type=str, required=True, default=None, help="The type of data", ) args = parser.parse_args() if args.data_type == "sft": seed = sft_seed elif args.data_type == "prompt": seed = prompt_seed elif args.data_type == "prompt_rlvr": seed = prompt_rlvr_seed elif args.data_type == "preference": seed = preference_seed elif args.data_type == "kto": seed = kto_seed else: raise ValueError(f"Unknown data type {args.data_type}") if args.data_type != "kto": line = json.dumps(seed, ensure_ascii=False) + "\n" for idx in [1, 2, 3]: with open(os.path.join(args.data_dir, f"{idx}.jsonl"), "w", encoding="utf8") as f: for i in range(1000): f.write(line) f.write(line) else: for idx in [1, 2, 3]: with open(os.path.join(args.data_dir, f"{idx}.jsonl"), "w", encoding="utf8") as f: for i in range(1000): seed["label"] = not seed["label"] line = json.dumps(seed, ensure_ascii=False) + "\n" f.write(line)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/tests/__init__.py
applications/ColossalChat/tests/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/tests/test_lora.py
applications/ColossalChat/tests/test_lora.py
import torch import torch.nn as nn import torch.optim as optim from coati.models import convert_to_lora_module from coati.models.lora import LoraConfig, LoraEmbedding, LoraLinear from torch.utils.data import DataLoader, TensorDataset class SimpleNN(nn.Module): def __init__(self, input_size, hidden_size, num_classes): super(SimpleNN, self).__init__() self.fc1 = nn.Linear(input_size, hidden_size) self.relu = nn.ReLU() self.fc2 = nn.Linear(hidden_size, num_classes) def forward(self, x): out = self.fc1(x) out = self.relu(out) out = self.fc2(out) return out def test_overfit(): input_size = 1000 hidden_size = 200 num_classes = 5 batch_size = 64 learning_rate = 0.01 num_epochs = 200 # Synthesized dataset X = torch.randn(batch_size, input_size) Y = torch.randint(0, num_classes, (batch_size,)) # Convert to DataLoader dataset = TensorDataset(X, Y) loader = DataLoader(dataset, batch_size=batch_size, shuffle=True) # Build and convert model model = SimpleNN(input_size, hidden_size, num_classes) weight_to_compare = model.fc1.weight.detach().clone() model = convert_to_lora_module(model, lora_config=LoraConfig(r=32)) # Loss and optimizer criterion = nn.CrossEntropyLoss() optimizer = optim.Adam(model.parameters(), lr=learning_rate) # Train the model for _ in range(num_epochs): for i, (inputs, labels) in enumerate(loader): # Forward pass outputs = model(inputs) loss = criterion(outputs, labels) # Backward and optimize optimizer.zero_grad() loss.backward() optimizer.step() # Check if model has overfitted outputs = model(X) _, predicted = torch.max(outputs.data, 1) total = labels.size(0) correct = (predicted == Y).sum().item() assert correct / total > 0.95 assert (weight_to_compare - model.fc1.weight).sum() < 0.01 def test_lora_linear_accuracy(): weight = torch.randn(10, 5) linear = nn.Linear(5, 10) linear.weight.data = weight x = torch.randn(10, 5) out_linear = linear(x) # lora linear Pissa linear.weight.data = weight lora_linear = LoraLinear(linear.weight, linear.bias, r=2, lora_initialization_method="PiSSA") out_lora = lora_linear(x) assert torch.allclose(out_linear, out_lora, atol=1e-5, rtol=1e-05) # lora linear linear.weight.data = weight lora_linear = LoraLinear(linear.weight, linear.bias, r=2) out_lora = lora_linear(x) assert torch.allclose(out_linear, out_lora, atol=1e-5, rtol=1e-05) def test_lora_embedding_accuracy(): weight = torch.randn(10, 5) embedding = nn.Embedding(10, 5) embedding.weight.data = weight x = torch.randint(0, 10, (10,)) out_embedding = embedding(x) # lora embedding Pissa embedding.weight.data = weight lora_embedding = LoraEmbedding( embedding.weight, r=2, lora_initialization_method="PiSSA", num_embeddings=10, embedding_dim=5 ) out_lora = lora_embedding(x) assert torch.allclose(out_embedding, out_lora, atol=1e-5, rtol=1e-05) # lora embedding embedding.weight.data = weight lora_embedding = LoraEmbedding(embedding.weight, r=2, num_embeddings=10, embedding_dim=5) out_lora = lora_embedding(x) assert torch.allclose(out_embedding, out_lora, atol=1e-5, rtol=1e-05) if __name__ == "__main__": test_overfit() test_lora_linear_accuracy() test_lora_embedding_accuracy()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/ray/train_prompts_on_ray.py
applications/ColossalChat/examples/community/ray/train_prompts_on_ray.py
import argparse import logging import os import socket from copy import deepcopy from typing import Type import ray import torch from coati.experience_maker.base import Experience from coati.models.base import RewardModel from coati.models.bloom import BLOOMActor, BLOOMCritic from coati.models.gpt import GPTActor, GPTCritic from coati.models.lora import LoRAModule from coati.models.loss import PolicyLoss, ValueLoss from coati.models.opt import OPTActor, OPTCritic from coati.models.utils import compute_reward from coati.trainer.strategies import DDPStrategy, GeminiStrategy, LowLevelZeroStrategy from ray.util.placement_group import placement_group from ray.util.scheduling_strategies import PlacementGroupSchedulingStrategy from torch.optim import Adam from transformers import AutoTokenizer, BloomTokenizerFast from transformers.models.gpt2.tokenization_gpt2 import GPT2Tokenizer from colossalai.nn.optimizer import HybridAdam class ExperienceCompositionRefs: def __init__( self, sequences_attention_mask_action_mask_ref: ray.ObjectRef, action_log_probs_ref: ray.ObjectRef, base_action_log_probs_ref: ray.ObjectRef, value_ref: ray.ObjectRef, r_ref: ray.ObjectRef, ) -> None: self.sequences_attention_mask_action_mask_ref = sequences_attention_mask_action_mask_ref self.action_log_probs_ref = action_log_probs_ref self.base_action_log_probs_ref = base_action_log_probs_ref self.value_ref = value_ref self.r_ref = r_ref class ExperienceMaker: def __init__(self, kl_coef) -> None: self.kl_coef = kl_coef @torch.no_grad() def make_experience(self, experiment_computation_refs: ExperienceCompositionRefs): sequences, attention_mask, action_mask = ray.get( experiment_computation_refs.sequences_attention_mask_action_mask_ref ) action_log_probs = ray.get(experiment_computation_refs.action_log_probs_ref) base_action_log_probs = ray.get(experiment_computation_refs.base_action_log_probs_ref) r = ray.get(experiment_computation_refs.r_ref) reward = compute_reward(r, self.kl_coef, action_log_probs, base_action_log_probs, action_mask=action_mask) value = ray.get(experiment_computation_refs.value_ref) advantage = reward - value if advantage.ndim == 1: advantage = advantage.unsqueeze(-1) experience = Experience(sequences, action_log_probs, value, reward, advantage, attention_mask, action_mask) return experience class DistributedTorchRayActor: def __init__(self, world_size, rank, local_rank, master_addr, master_port): logging.basicConfig( format="%(asctime)s %(levelname)-8s %(message)s", level=logging.INFO, datefmt="%Y-%m-%d %H:%M:%S" ) self._model = None self._world_size = world_size self._rank = rank self._local_rank = local_rank self._master_addr = master_addr if master_addr else self._get_current_node_ip() self._master_port = master_port if master_port else self._get_free_port() os.environ["MASTER_ADDR"] = self._master_addr os.environ["MASTER_PORT"] = str(self._master_port) os.environ["WORLD_SIZE"] = str(self._world_size) os.environ["RANK"] = str(self._rank) os.environ["LOCAL_RANK"] = str(self._local_rank) @staticmethod def _get_current_node_ip(): return ray._private.services.get_node_ip_address() @staticmethod def _get_free_port(): with socket.socket() as sock: sock.bind(("", 0)) return sock.getsockname()[1] def get_master_addr_port(self): return self._master_addr, self._master_port class BasePPORole(DistributedTorchRayActor): def add_experience_maker(self, kl_coef: float = 0.1): self._experience_maker = ExperienceMaker(kl_coef) def make_experience(self, experience_computation_ref: ExperienceCompositionRefs): return self._experience_maker.make_experience(experience_computation_ref) def _init_strategy(self, strategy: str): # configure strategy if strategy == "ddp": self._strategy = DDPStrategy() elif strategy == "colossalai_gemini": self._strategy = GeminiStrategy(placement_policy="cuda", initial_scale=2**5) elif strategy == "colossalai_zero2": self._strategy = LowLevelZeroStrategy(stage=2, placement_policy="cuda") else: raise ValueError(f'Unsupported strategy "{strategy}"') def _init_optimizer(self): if isinstance(self._strategy, (GeminiStrategy, LowLevelZeroStrategy)): self._optimizer = HybridAdam(self._model.parameters(), lr=5e-6) else: self._optimizer = Adam(self._model.parameters(), lr=5e-6) def _prepare_model_with_strategy(self, has_optimizer: bool): if has_optimizer: self._init_optimizer() (self._model, self._optimizer) = self._strategy.prepare((self._model, self._optimizer)) else: self._model = self._strategy.prepare(self._model) def _load_model_from_pretrained(self, model_class: Type[LoRAModule], pretrain: str): raise NotImplementedError() def init_model_from_pretrained( self, strategy: str, model_class: Type[LoRAModule], pretrain: str, has_optimizer=False ): self._init_strategy(strategy) self._load_model_from_pretrained(model_class, pretrain) self._prepare_model_with_strategy(has_optimizer) def eval(self): self._model.eval() class TrainablePPORole(BasePPORole): def _load_model_from_pretrained(self, model_class, pretrain): with self._strategy.model_init_context(): self._model = model_class(pretrain).to(torch.cuda.current_device()) def _train(self): self._model.train() def _training_step(self, experience: Experience): raise NotImplementedError() def learn_on_experiences(self, experience_refs): experiences = ray.get(experience_refs) device = torch.cuda.current_device() self._train() for exp in experiences: exp.to_device(device) self._training_step(exp) self.eval() @ray.remote(num_gpus=1) class RayPPOActor(TrainablePPORole): def set_loss_function(self, eps_clip: float): self._actor_loss_fn = PolicyLoss(eps_clip) def load_tokenizer_from_pretrained(self, model_type: str, pretrained): if model_type == "gpt2": self._model_tokenizer = GPT2Tokenizer.from_pretrained(pretrained) self._model_tokenizer.pad_token = self._model_tokenizer.eos_token elif model_type == "bloom": self._model_tokenizer = BloomTokenizerFast.from_pretrained(pretrained) self._model_tokenizer.pad_token = self._model_tokenizer.eos_token elif model_type == "opt": self._model_tokenizer = AutoTokenizer.from_pretrained(pretrained) else: raise ValueError(f'Unsupported model "{model_type}"') # Set tokenize function for sequence generation def _text_input_tokenize_fn(texts): batch = self._model_tokenizer(texts, return_tensors="pt", max_length=96, padding=True, truncation=True) return {k: v.cuda() for k, v in batch.items()} self._sample_tokenize_function = _text_input_tokenize_fn def setup_generate_kwargs(self, generate_kwargs: dict): from coati.trainer.ppo import _set_default_generate_kwargs self._generate_kwargs = _set_default_generate_kwargs(self._strategy, generate_kwargs, self._model) self._generate_kwargs["pad_token_id"] = self._model_tokenizer.pad_token_id self._generate_kwargs["eos_token_id"] = self._model_tokenizer.eos_token_id def load_csv_prompt_file_from_url_to_sampler(self, prompt_url): import pandas as pd prompts = pd.read_csv(prompt_url)["prompt"] self._sampler = self._strategy.setup_sampler(prompts) def _generate(self, input_ids, **generate_kwargs): return self._model.generate(input_ids, return_action_mask=True, **generate_kwargs) def sample_prompts_and_make_sequence(self, experience_batch_size): sampled_prompts = self._sampler.sample(experience_batch_size) input_ids = self._sample_tokenize_function(sampled_prompts) if isinstance(input_ids, dict): return self._generate(**input_ids, **self._generate_kwargs) else: return self._generate(input_ids, **self._generate_kwargs) @torch.no_grad() def calculate_action_log_probs(self, sequence_attention_action_mask): sequences, attention_mask, action_mask = sequence_attention_action_mask return self._model.forward(sequences, action_mask.size(1), attention_mask) def _training_step(self, experience): num_actions = experience.action_mask.size(1) action_log_probs = self._model(experience.sequences, num_actions, attention_mask=experience.attention_mask) actor_loss = self._actor_loss_fn( action_log_probs, experience.action_log_probs, experience.advantages, action_mask=experience.action_mask ) self._strategy.backward(actor_loss, self._model, self._optimizer) self._strategy.optimizer_step(self._optimizer) self._optimizer.zero_grad() logging.info("actor_loss: {}".format(actor_loss)) def save_checkpoint(self, save_path, should_save_optimizer: bool): if self._rank == 0: # save model checkpoint only on rank 0 self._strategy.save_model(self._model, save_path, only_rank0=True) # save optimizer checkpoint on all ranks if should_save_optimizer: self._strategy.save_optimizer( self._optimizer, "actor_optim_checkpoint_prompts_%d.pt" % (torch.cuda.current_device()), only_rank0=False, ) def generate_answer(self, prompt, max_length=30, num_return_sequences=5): encoded_input = self._model_tokenizer(prompt, return_tensors="pt") input_ids = {k: v.cuda() for k, v in encoded_input.items()} sequence, _ = self._model.generate( **input_ids, max_length=max_length, return_action_mask=False, num_return_sequences=num_return_sequences ) token_list = list(sequence.data[0]) output = " ".join([self._model_tokenizer.decode(token) for token in token_list]) return output @ray.remote(num_gpus=1) class RayPPOCritic(TrainablePPORole): def set_loss_function(self, value_clip: float): self._critic_loss_fn = ValueLoss(value_clip) def _training_step(self, experience): values = self._model( experience.sequences, action_mask=experience.action_mask, attention_mask=experience.attention_mask ) critic_loss = self._critic_loss_fn( values, experience.values, experience.reward, action_mask=experience.action_mask ) self._strategy.backward(critic_loss, self._model, self._optimizer) self._strategy.optimizer_step(self._optimizer) self._optimizer.zero_grad() logging.info("critic_loss: {}".format(critic_loss)) @torch.no_grad() def calculate_value(self, sequence_attention_action_mask): sequences, attention_mask, action_mask = sequence_attention_action_mask return self._model(sequences, action_mask, attention_mask) @ray.remote(num_gpus=1) class RayPPORewardModel(BasePPORole): def _load_model_from_pretrained(self, model_class, pretrain): with self._strategy.model_init_context(): critic = model_class(pretrained=pretrain).to(torch.cuda.current_device()) self._model = RewardModel(deepcopy(critic.model), deepcopy(critic.value_head)).to( torch.cuda.current_device() ) @torch.no_grad() def calculate_r(self, sequence_attention_action_mask): sequences, attention_mask, _ = sequence_attention_action_mask return self._model(sequences, attention_mask) @ray.remote(num_gpus=1) class RayPPOInitialModel(BasePPORole): def _load_model_from_pretrained(self, model_class, pretrain): with self._strategy.model_init_context(): self._model = model_class(pretrain).to(torch.cuda.current_device()) @torch.no_grad() def calculate_base_action_log_probs(self, sequence_attention_action_mask): sequences, attention_mask, action_mask = sequence_attention_action_mask return self._model(sequences, action_mask.size(1), attention_mask) class PPORayActorGroup: """ A group of ray actors Functions start with 'async' should return list of object refs """ def __init__(self, num_nodes, num_gpus_per_node, ray_actor_type: Type[BasePPORole]) -> None: self._num_nodes = num_nodes self._num_gpus_per_node = num_gpus_per_node self.ray_actor_type = ray_actor_type self._initiate_actors() def _initiate_actors(self): world_size = self._num_nodes * self._num_gpus_per_node # Use placement group to lock resources for models of same type pg = None if self._num_gpus_per_node > 1: bundles = [{"GPU": self._num_gpus_per_node, "CPU": self._num_gpus_per_node} for _ in range(self._num_nodes)] pg = placement_group(bundles, strategy="STRICT_SPREAD") ray.get(pg.ready()) if pg: master_actor = self.ray_actor_type.options( scheduling_strategy=PlacementGroupSchedulingStrategy(placement_group=pg, placement_group_bundle_index=0) ).remote(world_size, 0, 0, None, None) else: master_actor = self.ray_actor_type.options(num_gpus=1).remote(world_size, 0, 0, None, None) self._actor_handlers = [master_actor] # Create worker actors if world_size > 1: master_addr, master_port = ray.get(master_actor.get_master_addr_port.remote()) for rank in range(1, world_size): local_rank = rank % self._num_gpus_per_node if pg: worker_actor = self.ray_actor_type.options( scheduling_strategy=PlacementGroupSchedulingStrategy( placement_group=pg, placement_group_bundle_index=rank // self._num_gpus_per_node ) ).remote(world_size, rank, local_rank, master_addr, master_port) else: worker_actor = self.ray_actor_type.options(num_gpus=1).remote( world_size, rank, local_rank, master_addr, master_port ) self._actor_handlers.append(worker_actor) def async_init_model_from_pretrained( self, strategy: str, model_class: Type[LoRAModule], pretrain: str, has_optimizer: bool ): return [ actor.init_model_from_pretrained.remote(strategy, model_class, pretrain, has_optimizer) for actor in self._actor_handlers ] class TrainableModelRayActorGroup(PPORayActorGroup): def async_learn_on_experiences(self, experience_refs): num_actors = len(self._actor_handlers) learn_result_refs = [] for i in range(num_actors): exp_refs_batch = experience_refs[i::num_actors] learn_result_refs.append(self._actor_handlers[i].learn_on_experiences.remote(exp_refs_batch)) return learn_result_refs class PPOActorRayActorGroup(TrainableModelRayActorGroup): def __init__(self, num_nodes, num_gpus_per_node) -> None: super().__init__(num_nodes, num_gpus_per_node, RayPPOActor) def async_prepare_for_sequence_generation(self, model: str, pretrain: str, generation_kwargs: dict): refs = [] for actor in self._actor_handlers: refs.append(actor.load_tokenizer_from_pretrained.remote(model, pretrain)) refs.append(actor.setup_generate_kwargs.remote(generation_kwargs)) return refs def load_csv_prompt_file_from_url_to_sampler(self, csv_url): ray.get([actor.load_csv_prompt_file_from_url_to_sampler.remote(csv_url) for actor in self._actor_handlers]) def async_sample_prompts_and_make_sequence(self, experience_batch_size): return [actor.sample_prompts_and_make_sequence.remote(experience_batch_size) for actor in self._actor_handlers] def async_calculate_action_log_probs(self, sequences_attention_mask_action_mask_refs): num_actors = len(self._actor_handlers) action_log_probs_refs = [] for i in range(len(sequences_attention_mask_action_mask_refs)): action_log_probs_ref = self._actor_handlers[i % num_actors].calculate_action_log_probs.remote( sequences_attention_mask_action_mask_refs[i] ) action_log_probs_refs.append(action_log_probs_ref) return action_log_probs_refs def set_loss_function(self, eps_clip: float = 0.2): ray.get([actor.set_loss_function.remote(eps_clip) for actor in self._actor_handlers]) def save_checkpoint(self, save_path, should_save_optimizer): ray.get([actor.save_checkpoint.remote(save_path, should_save_optimizer) for actor in self._actor_handlers]) class PPOCriticRayActorGroup(TrainableModelRayActorGroup): def __init__(self, num_nodes, num_gpus_per_node) -> None: super().__init__(num_nodes, num_gpus_per_node, RayPPOCritic) def async_calculate_value(self, sequences_attention_mask_action_mask_refs): num_actors = len(self._actor_handlers) value_refs = [] for i in range(len(sequences_attention_mask_action_mask_refs)): value_ref = self._actor_handlers[i % num_actors].calculate_value.remote( sequences_attention_mask_action_mask_refs[i] ) value_refs.append(value_ref) return value_refs def set_loss_function(self, value_clip: float = 0.4): ray.get([actor.set_loss_function.remote(value_clip) for actor in self._actor_handlers]) class PPOInitialRayActorGroup(PPORayActorGroup): def __init__(self, num_nodes, num_gpus_per_node) -> None: super().__init__(num_nodes, num_gpus_per_node, RayPPOInitialModel) def async_calculate_base_action_log_probs(self, sequences_attention_mask_action_mask_refs): num_actors = len(self._actor_handlers) base_action_log_probs_refs = [] for i in range(len(sequences_attention_mask_action_mask_refs)): base_action_log_probs_ref = self._actor_handlers[i % num_actors].calculate_base_action_log_probs.remote( sequences_attention_mask_action_mask_refs[i] ) base_action_log_probs_refs.append(base_action_log_probs_ref) return base_action_log_probs_refs class PPORewardRayActorGroup(PPORayActorGroup): def __init__(self, num_nodes, num_gpus_per_node) -> None: super().__init__(num_nodes, num_gpus_per_node, RayPPORewardModel) def async_calculate_r(self, sequences_attention_mask_action_mask_refs): num_actors = len(self._actor_handlers) r_refs = [] for i in range(len(sequences_attention_mask_action_mask_refs)): r_ref = self._actor_handlers[i % num_actors].calculate_r.remote( sequences_attention_mask_action_mask_refs[i] ) r_refs.append(r_ref) return r_refs def main(args): logging.basicConfig( format="%(asctime)s %(levelname)-8s %(message)s", level=logging.INFO, datefmt="%Y-%m-%d %H:%M:%S" ) if args.model == "gpt2": actor_model_class, critic_model_class = GPTActor, GPTCritic elif args.model == "bloom": actor_model_class, critic_model_class = BLOOMActor, BLOOMCritic elif args.model == "opt": actor_model_class, critic_model_class = OPTActor, OPTCritic else: raise ValueError(f'Unsupported model "{args.model}"') logging.info("Start creating actors") # Initialize 4 models (actor, critic, initial_model and reward_model) actor_group = PPOActorRayActorGroup(num_nodes=args.num_actor_nodes, num_gpus_per_node=args.num_gpus_per_node) critic_group = PPOCriticRayActorGroup(num_nodes=args.num_critic_nodes, num_gpus_per_node=args.num_gpus_per_node) initial_group = PPOInitialRayActorGroup(num_nodes=args.num_initial_nodes, num_gpus_per_node=args.num_gpus_per_node) reward_group = PPORewardRayActorGroup(num_nodes=args.num_reward_nodes, num_gpus_per_node=args.num_gpus_per_node) logging.info("Actors created") # Prepare model for training generate_kwargs = {"max_length": 128, "do_sample": True, "temperature": 1.0, "top_k": 50} ray.get( actor_group.async_init_model_from_pretrained(args.strategy, actor_model_class, args.pretrain, True) + critic_group.async_init_model_from_pretrained(args.strategy, critic_model_class, args.pretrain, True) + initial_group.async_init_model_from_pretrained(args.strategy, actor_model_class, args.pretrain, False) + reward_group.async_init_model_from_pretrained(args.strategy, critic_model_class, args.pretrain, False) + actor_group.async_prepare_for_sequence_generation(args.model, args.pretrain, generate_kwargs) ) logging.info("Models prepared for training") # Prepare models for training actor_group.load_csv_prompt_file_from_url_to_sampler(args.prompt_csv_url) actor_group.set_loss_function() critic_group.set_loss_function() # Training parameter num_episodes = args.num_episodes max_timesteps = args.max_timesteps update_timesteps = args.update_timesteps experience_batch_size = args.experience_batch_size # Start training logging.info("Training start") # Set all models to eval and add experience maker all_ray_actors = ( actor_group._actor_handlers + critic_group._actor_handlers + initial_group._actor_handlers + reward_group._actor_handlers ) num_ray_actors = len(all_ray_actors) ray.get([ray_actor.eval.remote() for ray_actor in all_ray_actors]) ray.get([ray_actor.add_experience_maker.remote() for ray_actor in all_ray_actors]) # Used as a queue to coordinate experience making experience_composition_refs = [] time = 0 for episode in range(num_episodes): logging.info("episode {} started".format(episode)) for _ in range(max_timesteps): time += 1 # Experience queueing stage sequences_attention_mask_action_mask_refs = actor_group.async_sample_prompts_and_make_sequence( experience_batch_size ) base_action_log_probs_refs = initial_group.async_calculate_base_action_log_probs( sequences_attention_mask_action_mask_refs ) values_refs = critic_group.async_calculate_value(sequences_attention_mask_action_mask_refs) r_refs = reward_group.async_calculate_r(sequences_attention_mask_action_mask_refs) action_log_probs_refs = actor_group.async_calculate_action_log_probs( sequences_attention_mask_action_mask_refs ) experience_composition_refs.extend( [ ExperienceCompositionRefs( sequences_attention_mask_action_mask_refs[i], action_log_probs_refs[i], base_action_log_probs_refs[i], values_refs[i], r_refs[i], ) for i in range(len(sequences_attention_mask_action_mask_refs)) ] ) # Learning stage if time % update_timesteps == 0: experience_refs = [] # calculate experiences for i, experience_composition_ref in enumerate(experience_composition_refs): exp_composition_ref = experience_composition_ref selected_ray_actor = all_ray_actors[i % num_ray_actors] experience_refs.append(selected_ray_actor.make_experience.remote(exp_composition_ref)) # backward ray.get( actor_group.async_learn_on_experiences(experience_refs) + critic_group.async_learn_on_experiences(experience_refs) ) # clear refs queue experience_composition_refs.clear() logging.info("Training finished") # Save checkpoint actor_group.save_checkpoint(args.save_path, args.need_optim_ckpt) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_csv_url", type=str) parser.add_argument("--strategy", choices=["ddp", "colossalai_gemini", "colossalai_zero2"], default="ddp") parser.add_argument("--model", default="gpt2", choices=["gpt2", "bloom", "opt"]) parser.add_argument("--pretrain", type=str, default="gpt2") parser.add_argument("--save_path", type=str, default="actor_checkpoint_prompts.pt") parser.add_argument("--need_optim_ckpt", type=bool, default=False) parser.add_argument("--num_episodes", type=int, default=10) parser.add_argument("--max_timesteps", type=int, default=10) parser.add_argument("--update_timesteps", type=int, default=10) parser.add_argument("--train_batch_size", type=int, default=8) parser.add_argument("--experience_batch_size", type=int, default=8) parser.add_argument("--num_actor_nodes", type=int, help="num of nodes to use to host actor model", default=1) parser.add_argument("--num_critic_nodes", type=int, help="num of nodes to use to host critic model", default=1) parser.add_argument("--num_initial_nodes", type=int, help="num of nodes to use to host initial model", default=1) parser.add_argument("--num_reward_nodes", type=int, help="num of nodes to use to host reward model", default=1) parser.add_argument("--num_gpus_per_node", type=int, help="num of gpus on a ray node", default=1) args = parser.parse_args() ray.init() main(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/ray/ray_job_script.py
applications/ColossalChat/examples/community/ray/ray_job_script.py
import sys from ray.job_submission import JobSubmissionClient def main(api_server_endpoint="http://127.0.0.1:8265"): client = JobSubmissionClient(api_server_endpoint) client.submit_job( entrypoint="python experimental/ray/train_prompts_on_ray.py --strategy colossalai_zero2 --prompt_csv_url https://huggingface.co/datasets/fka/awesome-chatgpt-prompts/resolve/main/prompts.csv", runtime_env={ "working_dir": "applications/Chat", "pip": [ "torch==1.13.1", "transformers>=4.20.1", "datasets", "loralib", "colossalai>=0.2.4", "langchain", "tokenizers", "fastapi", "sse_starlette", "wandb", "sentencepiece", "gpustat", ], }, ) if __name__ == "__main__": main(sys.argv[1])
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/peft/train_peft_sft.py
applications/ColossalChat/examples/community/peft/train_peft_sft.py
import argparse import os import torch import torch.distributed as dist from coati.trainer import SFTTrainer from coati.trainer.strategies import DDPStrategy, GeminiStrategy, LowLevelZeroStrategy from easy_dataset import EasyDataset from peft import LoraConfig, PeftModel, TaskType, get_peft_model from torch.optim import Adam from torch.utils.data import DataLoader from torch.utils.data.dataloader import default_collate from torch.utils.data.distributed import DistributedSampler from transformers import AutoModelForCausalLM, AutoTokenizer, BloomTokenizerFast from transformers.models.gpt2.tokenization_gpt2 import GPT2Tokenizer from colossalai.logging import get_dist_logger from colossalai.nn.optimizer import HybridAdam from colossalai.tensor import ColoParameter def train(args): # configure strategy if args.strategy == "ddp": strategy = DDPStrategy() elif args.strategy == "colossalai_gemini": strategy = GeminiStrategy(placement_policy="static") elif args.strategy == "colossalai_zero2": strategy = LowLevelZeroStrategy(stage=2, placement_policy="cuda") else: raise ValueError(f'Unsupported strategy "{args.strategy}"') # configure model with strategy.model_init_context(): print("Warning: currently only bloom is tested, gpt2,llama and opt are not tested") model = AutoModelForCausalLM.from_pretrained(args.pretrain).to(torch.cuda.current_device()) # if the args.save_path exists and args.save_path+'/adapter_config.json' exists, we'll load the adapter_config.json if ( os.path.exists(args.save_path) and os.path.exists(args.save_path + "/adapter_config.json") and os.path.exists(args.save_path + "/adapter_model.bin") ): print("loading from saved peft model ", args.save_path) model = PeftModel.from_pretrained(model, args.save_path) else: # we'll use peft lora library to do the lora lora_rank = args.lora_rank if args.lora_rank > 0 else 32 # config lora with rank of lora_rank lora_config = LoraConfig( task_type=TaskType.CAUSAL_LM, inference_mode=False, r=lora_rank, lora_alpha=32, lora_dropout=0.1 ) model = get_peft_model(model, lora_config) model.print_trainable_parameters() # configure tokenizer if args.model == "gpt2": tokenizer = GPT2Tokenizer.from_pretrained("gpt2") tokenizer.pad_token = tokenizer.eos_token elif args.model == "bloom": tokenizer = BloomTokenizerFast.from_pretrained("bigscience/bloom-560m") tokenizer.pad_token = tokenizer.eos_token elif args.model == "opt": tokenizer = AutoTokenizer.from_pretrained("facebook/opt-350m") tokenizer.pad_token = tokenizer.eos_token elif args.model == "llama": tokenizer = AutoTokenizer.from_pretrained( args.pretrain, padding_side="right", use_fast=False, ) tokenizer.eos_token = "</s>" tokenizer.pad_token = tokenizer.unk_token else: raise ValueError(f'Unsupported model "{args.model}"') if args.model == "llama" and args.strategy == "colossalai_gemini": # this is a hack to deal with the resized embedding # to make sure all parameters are ColoParameter for Colossal-AI Gemini Compatibility for name, param in model.named_parameters(): if not isinstance(param, ColoParameter): sub_module_name = ".".join(name.split(".")[:-1]) weight_name = name.split(".")[-1] sub_module = model.get_submodule(sub_module_name) setattr(sub_module, weight_name, ColoParameter(param)) # configure optimizer if args.strategy.startswith("colossalai"): optim = HybridAdam(model.parameters(), lr=args.lr, clipping_norm=1.0) else: optim = Adam(model.parameters(), lr=args.lr) logger = get_dist_logger() logger.set_level("WARNING") # configure dataset law_dataset = EasyDataset(args.dataset, tokenizer=tokenizer, is_group_texts=not args.is_short_text) train_dataset = law_dataset print(train_dataset) eval_dataset = None if args.eval_dataset is not None: eval_dataset = EasyDataset(args.eval_dataset, tokenizer=tokenizer, is_group_texts=not args.is_short_text) data_collator = default_collate if dist.is_initialized() and dist.get_world_size() > 1: train_sampler = DistributedSampler( train_dataset, shuffle=True, seed=42, drop_last=True, rank=dist.get_rank(), num_replicas=dist.get_world_size(), ) if eval_dataset is not None: eval_sampler = DistributedSampler( eval_dataset, shuffle=False, seed=42, drop_last=False, rank=dist.get_rank(), num_replicas=dist.get_world_size(), ) else: train_sampler = None eval_sampler = None train_dataloader = DataLoader( train_dataset, shuffle=(train_sampler is None), sampler=train_sampler, batch_size=args.batch_size, collate_fn=data_collator, pin_memory=True, ) if eval_dataset is not None: eval_dataloader = DataLoader( eval_dataset, shuffle=(eval_sampler is None), sampler=eval_sampler, batch_size=args.batch_size, collate_fn=data_collator, pin_memory=True, ) else: eval_dataloader = None trainer = SFTTrainer( model=model, strategy=strategy, optim=optim, train_dataloader=train_dataloader, eval_dataloader=eval_dataloader, batch_size=args.batch_size, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, ) trainer.fit(logger=logger, log_interval=args.log_interval) # save model checkpoint after fitting on only rank0 trainer.save_model(path=args.save_path, only_rank0=True, tokenizer=tokenizer) # save optimizer checkpoint on all ranks if args.need_optim_ckpt: strategy.save_optimizer( trainer.optimizer, "rm_optim_checkpoint_%d.pt" % (torch.cuda.current_device()), only_rank0=False ) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--strategy", choices=["ddp", "colossalai_gemini", "colossalai_zero2"], default="ddp") parser.add_argument("--model", choices=["gpt2", "bloom", "opt", "llama"], default="bloom") parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--dataset", type=str, default=None) parser.add_argument("--eval_dataset", type=str, default=None) parser.add_argument("--save_path", type=str, default="output") parser.add_argument("--need_optim_ckpt", type=bool, default=False) parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--lora_rank", type=int, default=0, help="low-rank adaptation matrices rank") parser.add_argument("--log_interval", type=int, default=100, help="how many steps to log") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--enable_peft_lora", action="store_true", default=False) parser.add_argument("--is_short_text", action="store_true", default=False) args = parser.parse_args() train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/peft/easy_dataset.py
applications/ColossalChat/examples/community/peft/easy_dataset.py
import copy import json from typing import Dict, Sequence import torch from torch.utils.data import Dataset from tqdm import tqdm from transformers import AutoTokenizer IGNORE_INDEX = -100 def _tokenize_fn(strings: Sequence[str], tokenizer: AutoTokenizer, max_length: int = 512) -> Dict: """Tokenize a list of strings.""" tokenized_list = [ tokenizer( text, return_tensors="pt", padding="longest", max_length=max_length, truncation=True, ) for text in strings ] input_ids = labels = [tokenized.input_ids[0] for tokenized in tokenized_list] input_ids_lens = labels_lens = [ tokenized.input_ids.ne(tokenizer.pad_token_id).sum().item() for tokenized in tokenized_list ] return dict( input_ids=input_ids, labels=labels, input_ids_lens=input_ids_lens, labels_lens=labels_lens, ) def preprocess(sources: Sequence[str], targets: Sequence[str], tokenizer: AutoTokenizer, max_length: int = 512) -> Dict: """Preprocess the data by tokenizing.""" examples = [s + t for s, t in zip(sources, targets)] examples_tokenized, sources_tokenized = [ _tokenize_fn(strings, tokenizer, max_length) for strings in (examples, sources) ] input_ids = examples_tokenized["input_ids"] labels = copy.deepcopy(input_ids) for label, source_len in zip(labels, sources_tokenized["input_ids_lens"]): label[:source_len] = IGNORE_INDEX return dict(input_ids=input_ids, labels=labels) class EasySupervisedDataset(Dataset): def __init__(self, data_file: str, tokenizer: AutoTokenizer, max_length: int = 512) -> None: super(EasySupervisedDataset, self).__init__() with open(data_file, "r", encoding="UTF-8") as f: all_lines = f.readlines() # split to source and target ,source the characters before "回答:" including "回答:", target the characters after "回答:" sources, targets = [], [] for line in all_lines: if "回答:" in line: sep_index = line.index("回答:") sources.append(line[: sep_index + 3]) targets.append(line[sep_index + 3 :] + tokenizer.eos_token) else: sources.append(line) targets.append("" + tokenizer.eos_token) data_dict = preprocess(sources, targets, tokenizer, max_length) self.input_ids = data_dict["input_ids"] self.labels = data_dict["labels"] self.data_file = data_file def __len__(self): return len(self.input_ids) def __getitem__(self, i) -> Dict[str, torch.Tensor]: return dict(input_ids=self.input_ids[i], labels=self.labels[i]) def __repr__(self): return f"LawSupervisedDataset(data_file={self.data_file}, input_ids_len={len(self.input_ids)}, labels_len={len(self.labels)})" def __str__(self): return f"LawSupervisedDataset(data_file={self.data_file}, input_ids_len={len(self.input_ids)}, labels_len={len(self.labels)})" class EasyPromptsDataset(Dataset): def __init__(self, data_file: str, tokenizer: AutoTokenizer, max_length: int = 96) -> None: super(EasyPromptsDataset, self).__init__() with open(data_file, "r", encoding="UTF-8") as f: all_lines = f.readlines() all_lines = [line if "回答:" not in line else line[: line.index("回答:") + 3] for line in all_lines] self.prompts = [ tokenizer(line, return_tensors="pt", max_length=max_length, padding="max_length", truncation=True)[ "input_ids" ] .to(torch.cuda.current_device()) .squeeze(0) for line in tqdm(all_lines) ] self.data_file = data_file def __len__(self): return len(self.prompts) def __getitem__(self, idx): return self.prompts[idx] def __repr__(self): return f"LawPromptsDataset(data_file={self.data_file}, prompts_len={len(self.prompts)})" def __str__(self): return f"LawPromptsDataset(data_file={self.data_file}, prompts_len={len(self.prompts)})" class EasyRewardDataset(Dataset): def __init__(self, train_file: str, tokenizer: AutoTokenizer, special_token=None, max_length=512) -> None: super(EasyRewardDataset, self).__init__() self.chosen = [] self.reject = [] if special_token is None: self.end_token = tokenizer.eos_token else: self.end_token = special_token print(self.end_token) # read all lines in the train_file to a list with open(train_file, "r", encoding="UTF-8") as f: all_lines = f.readlines() for line in tqdm(all_lines): data = json.loads(line) prompt = "提问:" + data["prompt"] + " 回答:" chosen = prompt + data["chosen"] + self.end_token chosen_token = tokenizer( chosen, max_length=max_length, padding="max_length", truncation=True, return_tensors="pt" ) self.chosen.append( {"input_ids": chosen_token["input_ids"], "attention_mask": chosen_token["attention_mask"]} ) reject = prompt + data["rejected"] + self.end_token reject_token = tokenizer( reject, max_length=max_length, padding="max_length", truncation=True, return_tensors="pt" ) self.reject.append( {"input_ids": reject_token["input_ids"], "attention_mask": reject_token["attention_mask"]} ) def __len__(self): length = len(self.chosen) return length def __getitem__(self, idx): return ( self.chosen[idx]["input_ids"], self.chosen[idx]["attention_mask"], self.reject[idx]["input_ids"], self.reject[idx]["attention_mask"], ) # python representation of the object and the string representation of the object def __repr__(self): return f"LawRewardDataset(chosen_len={len(self.chosen)}, reject_len={len(self.reject)})" def __str__(self): return f"LawRewardDataset(chosen_len={len(self.chosen)}, reject_len={len(self.reject)})" """ Easy SFT just accept a text file which can be read line by line. However the datasets will group texts together to max_length so LLM will learn the texts meaning better. If individual lines are not related, just set is_group_texts to False. """ class EasySFTDataset(Dataset): def __init__(self, data_file: str, tokenizer: AutoTokenizer, max_length=512, is_group_texts=True) -> None: super().__init__() # read the data_file line by line with open(data_file, "r", encoding="UTF-8") as f: # encode the text data line by line and put raw python list input_ids only to raw_input_ids list raw_input_ids = [] for line in f: encoded_ids = tokenizer.encode(line) # if the encoded_ids is longer than max_length, then split it into several parts if len(encoded_ids) > max_length: for i in range(0, len(encoded_ids), max_length): raw_input_ids.append(encoded_ids[i : i + max_length]) else: raw_input_ids.append(encoded_ids) grouped_input_ids = [] current_input_ids = [] attention_mask = [] if tokenizer.pad_token_id is None: tokenizer.pad_token_id = tokenizer.eos_token_id if is_group_texts: for input_ids in raw_input_ids: if len(current_input_ids) + len(input_ids) > max_length: # pad the current_input_ids to max_length with tokenizer.pad_token_id padded_length = max_length - len(current_input_ids) current_input_ids.extend([tokenizer.pad_token_id] * padded_length) grouped_input_ids.append(torch.tensor(current_input_ids, dtype=torch.long)) attention_mask.append( torch.tensor([1] * (max_length - padded_length) + [0] * padded_length, dtype=torch.long) ) current_input_ids = [] else: current_input_ids.extend(input_ids) if len(current_input_ids) > 0: padded_length = max_length - len(current_input_ids) current_input_ids.extend([tokenizer.pad_token_id] * padded_length) grouped_input_ids.append(torch.tensor(current_input_ids, dtype=torch.long)) attention_mask.append( torch.tensor([1] * (max_length - padded_length) + [0] * padded_length, dtype=torch.long) ) else: # just append the raw_input_ids to max_length for input_ids in raw_input_ids: padded_length = max_length - len(input_ids) input_ids.extend([tokenizer.pad_token_id] * padded_length) attention_mask.append( torch.tensor([1] * (max_length - padded_length) + [0] * padded_length, dtype=torch.long) ) grouped_input_ids.append(torch.tensor(input_ids, dtype=torch.long)) self.input_ids = grouped_input_ids self.labels = copy.deepcopy(self.input_ids) self.file_name = data_file self.attention_mask = attention_mask def __len__(self): return len(self.input_ids) # get item from dataset def __getitem__(self, idx): return dict(input_ids=self.input_ids[idx], labels=self.labels[idx], attention_mask=self.attention_mask[idx]) # generate the dataset description to be printed by print in python def __repr__(self): return f"EasySFTDataset(len={len(self)},\nfile_name is {self.file_name})" # generate the dataset description to be printed by print in python def __str__(self): return f"EasySFTDataset(len={len(self)},\nfile_name is {self.file_name})"
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/peft/train_peft_prompts.py
applications/ColossalChat/examples/community/peft/train_peft_prompts.py
import argparse import torch import torch.distributed as dist from coati.dataset import DataCollatorForSupervisedDataset from coati.models.bloom import BLOOMRM, BLOOMCritic from coati.models.gpt import GPTRM, GPTCritic from coati.models.llama import LlamaCritic, LlamaRM from coati.models.opt import OPTRM, OPTCritic from coati.trainer import PPOTrainer from coati.trainer.strategies import DDPStrategy, GeminiStrategy, LowLevelZeroStrategy from easy_dataset import EasyPromptsDataset, EasySupervisedDataset from easy_models import BLOOMActor from torch.optim import Adam from torch.utils.data import DataLoader from torch.utils.data.distributed import DistributedSampler from transformers import AutoTokenizer, BloomTokenizerFast, GPT2Tokenizer, LlamaTokenizer from colossalai.nn.optimizer import HybridAdam def main(args): # configure strategy if args.strategy == "ddp": strategy = DDPStrategy() elif args.strategy == "colossalai_gemini": strategy = GeminiStrategy( placement_policy="static", offload_optim_frac=1.0, offload_param_frac=1.0, initial_scale=2**5 ) elif args.strategy == "colossalai_zero2": strategy = LowLevelZeroStrategy(stage=2, placement_policy="cpu") else: raise ValueError(f'Unsupported strategy "{args.strategy}"') if args.rm_path is not None: state_dict = torch.load(args.rm_path, map_location="cpu") # configure model if args.model == "bloom": # initial_model = BLOOMActor(pretrained=args.pretrain) print("Using peft lora to load Bloom model as initial_model") initial_model = BLOOMActor(pretrained=args.pretrain, lora_path=args.sft_lora_path) print("Using peft lora to load Bloom model as initial_model (Done)") else: raise ValueError(f'Unsupported actor model "{args.model}"') if args.rm_model == None: rm_model_name = args.model else: rm_model_name = args.rm_model if rm_model_name == "gpt2": reward_model = GPTRM(pretrained=args.rm_pretrain) elif rm_model_name == "bloom": print("load bloom reward model ", args.rm_pretrain) reward_model = BLOOMRM(pretrained=args.rm_pretrain) elif rm_model_name == "opt": reward_model = OPTRM(pretrained=args.rm_pretrain) elif rm_model_name == "llama": reward_model = LlamaRM(pretrained=args.rm_pretrain) else: raise ValueError(f'Unsupported reward model "{rm_model_name}"') if args.rm_path is not None: print("Loading reward model from", args.rm_path) reward_model.load_state_dict(state_dict) if args.strategy != "colossalai_gemini": initial_model.to(torch.float16).to(torch.cuda.current_device()) reward_model.to(torch.float16).to(torch.cuda.current_device()) with strategy.model_init_context(): if args.model == "bloom": # actor = BLOOMActor(pretrained=args.pretrain, lora_rank=args.lora_rank) print("Using peft lora to load Bloom model as Actor") actor = BLOOMActor(pretrained=args.pretrain, lora_path=args.sft_lora_path) print("Using peft lora to load Bloom model as Actor (Done)") else: raise ValueError(f'Unsupported actor model "{args.model}"') if rm_model_name == "gpt2": critic = GPTCritic(pretrained=args.rm_pretrain, lora_rank=args.lora_rank, use_action_mask=True) elif rm_model_name == "bloom": print("load bloom critic ", args.rm_pretrain, " lora_rank ", args.lora_rank, " use_action_mask ", True) critic = BLOOMCritic(pretrained=args.rm_pretrain, lora_rank=args.lora_rank, use_action_mask=True) print("load bloom critic (Done) ") elif rm_model_name == "opt": critic = OPTCritic(pretrained=args.rm_pretrain, lora_rank=args.lora_rank, use_action_mask=True) elif rm_model_name == "llama": critic = LlamaCritic(pretrained=args.rm_pretrain, lora_rank=args.lora_rank, use_action_mask=True) else: raise ValueError(f'Unsupported reward model "{rm_model_name}"') if args.rm_path is not None: print("Loading reward model from", args.rm_path) critic.load_state_dict(state_dict) del state_dict if args.strategy != "colossalai_gemini": critic.to(torch.float16).to(torch.cuda.current_device()) actor.to(torch.float16).to(torch.cuda.current_device()) # configure optimizer if args.strategy.startswith("colossalai"): actor_optim = HybridAdam(actor.parameters(), lr=1e-7) critic_optim = HybridAdam(critic.parameters(), lr=1e-7) else: actor_optim = Adam(actor.parameters(), lr=1e-7) critic_optim = Adam(critic.parameters(), lr=1e-7) # configure tokenizer if args.model == "gpt2": tokenizer = GPT2Tokenizer.from_pretrained(args.rm_pretrain) tokenizer.pad_token = tokenizer.eos_token elif args.model == "bloom": tokenizer = BloomTokenizerFast.from_pretrained(args.rm_pretrain) tokenizer.pad_token = tokenizer.eos_token elif args.model == "opt": tokenizer = AutoTokenizer.from_pretrained(args.rm_pretrain) tokenizer.pad_token = tokenizer.eos_token elif args.model == "llama": tokenizer = LlamaTokenizer.from_pretrained(args.pretrain) tokenizer.eos_token = "</s>" tokenizer.pad_token = tokenizer.unk_token else: raise ValueError(f'Unsupported model "{args.model}"') data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer) prompt_dataset = EasyPromptsDataset(args.prompt_path, tokenizer) if dist.is_initialized() and dist.get_world_size() > 1: prompt_sampler = DistributedSampler(prompt_dataset, shuffle=True, seed=42, drop_last=True) else: prompt_sampler = None prompt_dataloader = DataLoader( prompt_dataset, shuffle=(prompt_sampler is None), sampler=prompt_sampler, batch_size=args.train_batch_size ) pretrain_dataset = EasySupervisedDataset(args.pretrain_dataset, tokenizer) if dist.is_initialized() and dist.get_world_size() > 1: pretrain_sampler = DistributedSampler(pretrain_dataset, shuffle=True, seed=42, drop_last=True) else: pretrain_sampler = None pretrain_dataloader = DataLoader( pretrain_dataset, shuffle=(pretrain_sampler is None), sampler=pretrain_sampler, batch_size=args.ptx_batch_size, collate_fn=data_collator, ) def tokenize_fn(texts): # MUST padding to max length to ensure inputs of all ranks have the same length # Different length may lead to hang when using gemini, as different generation steps batch = tokenizer(texts, return_tensors="pt", max_length=96, padding="max_length", truncation=True) return {k: v.to(torch.cuda.current_device()) for k, v in batch.items()} (actor, actor_optim), (critic, critic_optim) = strategy.prepare((actor, actor_optim), (critic, critic_optim)) # configure trainer trainer = PPOTrainer( strategy, actor, critic, reward_model, initial_model, actor_optim, critic_optim, kl_coef=args.kl_coef, ptx_coef=args.ptx_coef, train_batch_size=args.train_batch_size, experience_batch_size=args.experience_batch_size, tokenizer=tokenize_fn, max_length=512, do_sample=True, temperature=1.0, top_k=50, pad_token_id=tokenizer.pad_token_id, eos_token_id=tokenizer.eos_token_id, ) trainer.fit( prompt_dataloader=prompt_dataloader, pretrain_dataloader=pretrain_dataloader, num_episodes=args.num_episodes, num_update_steps=args.num_update_steps, num_collect_steps=args.num_collect_steps, ) # save model checkpoint after fitting trainer.save_model(args.save_path, only_rank0=True, tokenizer=tokenizer) # save optimizer checkpoint on all ranks if args.need_optim_ckpt: strategy.save_optimizer( actor_optim, "actor_optim_checkpoint_prompts_%d.pt" % (torch.cuda.current_device()), only_rank0=False ) if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_path", type=str, default=None, help="path to the prompt dataset") parser.add_argument("--pretrain_dataset", type=str, default=None, help="path to the pretrained dataset") parser.add_argument( "--strategy", choices=["ddp", "colossalai_gemini", "colossalai_zero2"], default="ddp", help="strategy to use" ) parser.add_argument("--model", default="gpt2", choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--sft_lora_path", type=str, default=None) parser.add_argument("--rm_model", default=None, choices=["gpt2", "bloom", "opt", "llama"]) parser.add_argument("--rm_path", type=str, default=None) parser.add_argument("--rm_pretrain", type=str, default=None) parser.add_argument("--save_path", type=str, default="actor_checkpoint_prompts") parser.add_argument("--need_optim_ckpt", type=bool, default=False) parser.add_argument("--num_episodes", type=int, default=10) parser.add_argument("--num_collect_steps", type=int, default=10) parser.add_argument("--num_update_steps", type=int, default=5) parser.add_argument("--train_batch_size", type=int, default=2) parser.add_argument("--ptx_batch_size", type=int, default=1) parser.add_argument("--experience_batch_size", type=int, default=8) parser.add_argument("--lora_rank", type=int, default=0, help="low-rank adaptation matrices rank") parser.add_argument("--kl_coef", type=float, default=0.1) parser.add_argument("--ptx_coef", type=float, default=0.9) args = parser.parse_args() main(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/community/peft/easy_models.py
applications/ColossalChat/examples/community/peft/easy_models.py
from typing import Optional, Tuple, Union import torch import torch.nn as nn import torch.nn.functional as F from coati.models.generation import generate from coati.models.utils import log_probs_from_logits from peft import PeftModel from torch.nn.modules import Module from transformers import BloomConfig, BloomForCausalLM class Actor(Module): """ Actor model base class. Args: model (nn.Module): Actor Model. """ def __init__(self, model: nn.Module) -> None: super().__init__() self.model = model @torch.no_grad() def generate( self, input_ids: torch.Tensor, return_action_mask: bool = True, **kwargs ) -> Union[Tuple[torch.LongTensor, torch.LongTensor], Tuple[torch.LongTensor, torch.LongTensor, torch.BoolTensor]]: sequences = generate(self.model, input_ids, **kwargs) attention_mask = None pad_token_id = kwargs.get("pad_token_id", None) if pad_token_id is not None: attention_mask = sequences.not_equal(pad_token_id).to(dtype=torch.long, device=sequences.device) if not return_action_mask: return sequences, attention_mask, None input_len = input_ids.size(1) eos_token_id = kwargs.get("eos_token_id", None) if eos_token_id is None: action_mask = torch.ones_like(sequences, dtype=torch.bool) else: # left padding may be applied, only mask action action_mask = (sequences[:, input_len:] == eos_token_id).cumsum(dim=-1) == 0 action_mask = F.pad(action_mask, (1 + input_len, -1), value=True) # include eos token and input action_mask[:, :input_len] = False action_mask = action_mask[:, 1:] return sequences, attention_mask, action_mask[:, -(sequences.size(1) - input_len) :] def forward( self, sequences: torch.LongTensor, num_actions: int, attention_mask: Optional[torch.Tensor] = None ) -> torch.Tensor: """Returns action log probs""" output = self.model(sequences, attention_mask=attention_mask) logits = output["logits"] log_probs = log_probs_from_logits(logits[:, :-1, :], sequences[:, 1:]) return log_probs[:, -num_actions:] def get_base_model(self): return self.model class BLOOMActor(Actor): """ BLOOM Actor model. Args: pretrained (str): Pretrained model name or path. config (BloomConfig): Model config. checkpoint (bool): Enable gradient checkpointing. lora_rank (int): LoRA rank. lora_train_bias (str): LoRA bias training mode. """ def __init__( self, pretrained: str = None, config: Optional[BloomConfig] = None, checkpoint: bool = False, lora_path: str = None, ) -> None: if pretrained is not None: model = BloomForCausalLM.from_pretrained(pretrained) elif config is not None: model = BloomForCausalLM(config) else: model = BloomForCausalLM(BloomConfig()) if lora_path is not None: model = PeftModel.from_pretrained(model, lora_path) if checkpoint: model.gradient_checkpointing_enable() super().__init__(model) def print_trainable_parameters(self): self.get_base_model().print_trainable_parameters()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_ppo.py
applications/ColossalChat/examples/training_scripts/train_ppo.py
import argparse import json import os import resource from contextlib import nullcontext import torch import torch.distributed as dist from coati.dataset import ( DataCollatorForPromptDataset, DataCollatorForSupervisedDataset, StatefulDistributedSampler, load_tokenized_dataset, setup_conversation_template, ) from coati.models import ( Critic, LoraConfig, RewardModel, RLVRRewardModel, convert_to_lora_module, disable_dropout, lora_manager, ) from coati.trainer import PPOTrainer from coati.utils import load_checkpoint from coati.utils.reward_score import * from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam from colossalai.shardformer.policies.auto_policy import get_autopolicy logger = get_dist_logger() # default settings for response format tags, overwrite it in chat_template definition if needed response_format_tags = { "think_start": {"text": "<think>", "num_occur": 1}, "think_end": {"text": "</think>", "num_occur": 1}, "answer_start": {"text": "<answer>", "num_occur": 1}, "answer_end": {"text": "</answer>", "num_occur": 1}, } def train(args): global response_format_tags lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: actor = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) ref_model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) if not args.no_neural_reward_model: reward_model = RewardModel( args.rm_pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) critic = Critic( args.rm_pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True) ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True) if not args.no_neural_reward_model: reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True) critic = Critic(args.rm_pretrain) if args.lora_config is not None: actor = convert_to_lora_module(actor, lora_config=lora_config) critic = convert_to_lora_module(critic, lora_config=lora_config) for name, module in actor.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) for name, module in critic.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) lora_manager.able_to_merge = False # Disable dropout disable_dropout(actor) disable_dropout(critic) if args.grad_checkpoint: actor.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) critic.model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if os.path.exists(args.conversation_template_config): with open(args.conversation_template_config, "r", encoding="utf8") as f: conversation_template_config = json.load(f) dist.barrier() if "response_format_tags" in conversation_template_config: logger.warning(f"Overwrite default response format tags with {args.conversation_template_config}") response_format_tags = conversation_template_config.get("response_format_tags", response_format_tags) conversation_template = setup_conversation_template( tokenizer, chat_template_config=conversation_template_config, save_path=args.conversation_template_config ) stop_ids = conversation_template.stop_ids if len(conversation_template.stop_ids) > 0 else None else: raise ValueError("Conversation template config is not provided or incorrect") if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False tokenizer.padding_side = "left" # left padding for generation (online learning) # configure generation config actor.generation_config.update( pad_token_id=tokenizer.eos_token_id, bos_token_id=tokenizer.bos_token_id, eos_token_id=tokenizer.eos_token_id ) # configure optimizer coordinator.print_on_master(f"setting up optimizer for actor: lr={args.lr}, weight_decay={args.weight_decay}") actor_optim = HybridAdam( model_params=actor.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) coordinator.print_on_master(f"setting up optimizer for critic: lr={args.lr}, weight_decay={args.weight_decay}") critic_optim = HybridAdam( model_params=critic.parameters(), lr=args.critic_lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) if args.warmup_steps is None: args.warmup_steps = int(0.025 * args.num_episodes) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") actor_lr_scheduler = CosineAnnealingWarmupLR( optimizer=actor_optim, total_steps=args.num_episodes, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) critic_lr_scheduler = CosineAnnealingWarmupLR( optimizer=critic_optim, total_steps=args.num_episodes, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=True) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": if args.use_flash_attn and (args.tp > 1 or args.pp > 1 or args.sp > 1 or args.enable_sequence_parallelism): logger.warning("Flash attention cannot be used with 3D parallelism for PPO training. Disabling it.") args.use_flash_attn = False plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, ) custom_plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, custom_policy=get_autopolicy(critic.model), ) else: raise ValueError(f"Unknown plugin {args.plugin}") if args.plugin != "3d": custom_plugin = plugin # configure dataset coordinator.print_on_master(f"Load dataset: {args.prompt_dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_prompt_dataset = load_tokenized_dataset(dataset_paths=args.prompt_dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForPromptDataset(tokenizer=tokenizer, max_length=args.max_length - args.max_seq_len) train_prompt_dataloader = plugin.prepare_dataloader( dataset=train_prompt_dataset, batch_size=args.experience_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) if len(args.ptx_dataset) > 0: train_ptx_dataset = load_tokenized_dataset(dataset_paths=args.ptx_dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer, max_length=args.max_length) train_pretrain_dataloader = plugin.prepare_dataloader( dataset=train_ptx_dataset, batch_size=args.ptx_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: train_pretrain_dataloader = None actor_booster = Booster(plugin=plugin) ref_booster = Booster(plugin=plugin) if not args.no_neural_reward_model: rm_booster = Booster(plugin=custom_plugin) critic_booster = Booster(plugin=custom_plugin) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) actor, actor_optim, _, train_prompt_dataloader, actor_lr_scheduler = actor_booster.boost( model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, dataloader=train_prompt_dataloader, ) critic, critic_optim, _, _, critic_lr_scheduler = critic_booster.boost( model=critic, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, dataloader=train_prompt_dataloader, ) if not args.no_neural_reward_model: reward_model, _, _, _, _ = rm_booster.boost(model=reward_model, dataloader=train_prompt_dataloader) else: if args.reward_functions: reward_fn_list = [] for reward_fn in args.reward_functions: """ To define custom reward function, you can define your functions under: colossalai/applications/ColossalChat/coati/utils/reward_score/__init__.py and use it here by mofiying the following line: """ if reward_fn == "gsm8k_reward_fn": reward_fn_list.append(gsm8k_reward_fn) elif reward_fn == "math_competition_reward_fn": reward_fn_list.append(math_competition_reward_fn) else: raise ValueError(f"Unknown reward function {reward_fn}") reward_fn_list.append(eval(reward_fn)) reward_model = RLVRRewardModel( reward_fn_list=reward_fn_list, tokenizer=tokenizer, tags=response_format_tags ) ref_model, _, _, _, _ = ref_booster.boost(model=ref_model, dataloader=train_prompt_dataloader) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) sampler_start_idx = 0 start_step = 0 if args.rm_checkpoint_path is not None: if "modeling" in args.rm_checkpoint_path: rm_booster.load_model(reward_model, args.rm_checkpoint_path) else: _, _, _ = load_checkpoint( load_dir=args.rm_checkpoint_path, booster=rm_booster, model=reward_model, optimizer=None, lr_scheduler=None, ) coordinator.print_on_master(f"Loaded reward model checkpoint {args.rm_checkpoint_path}") if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: actor_booster.load_model(actor, args.checkpoint_path) ref_booster.load_model(ref_model, args.checkpoint_path) coordinator.print_on_master(f"Loaded actor and reference model {args.checkpoint_path}") else: _, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=actor_booster, model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, ) _, _, _ = load_checkpoint( load_dir=args.checkpoint_path, booster=ref_booster, model=ref_model, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, ) assert isinstance(train_prompt_dataloader.sampler, StatefulDistributedSampler) train_prompt_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded actor and reference model checkpoint {args.checkpoint_path} at spisode {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) if args.critic_checkpoint_path is not None: if "modeling" in args.critic_checkpoint_path: critic_booster.load_model(critic, args.critic_checkpoint_path) else: _, _, _ = load_checkpoint( load_dir=args.critic_checkpoint_path, booster=critic_booster, model=critic, optimizer=critic_optim, lr_scheduler=critic_lr_scheduler, ) coordinator.print_on_master(f"Loaded critic checkpoint {args.critic_checkpoint_path}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) # configure trainer trainer = PPOTrainer( actor_booster, critic_booster, actor, critic, reward_model, ref_model, actor_optim, critic_optim, actor_lr_scheduler, critic_lr_scheduler, tokenizer=tokenizer, stop_token_ids=stop_ids, kl_coef=args.kl_coef, ptx_coef=args.ptx_coef, train_batch_size=args.train_batch_size, buffer_limit=args.num_collect_steps * args.experience_batch_size, max_length=args.max_length, max_new_tokens=args.max_seq_len, use_cache=True, do_sample=True, temperature=0.7, apply_loss_mask=not args.disable_loss_mask, accumulation_steps=args.accumulation_steps, save_dir=args.save_path, save_interval=args.save_interval, top_k=50, use_tp=args.tp > 1, offload_inference_models="gemini" not in args.plugin, coordinator=coordinator, ) trainer.fit( num_episodes=args.num_episodes, num_collect_steps=args.num_collect_steps, num_update_steps=args.num_update_steps, prompt_dataloader=train_prompt_dataloader, pretrain_dataloader=train_pretrain_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights lora_manager.able_to_merge = True actor.eval() critic.eval() # save model checkpoint after fitting on only rank0 coordinator.print_on_master("Start saving final actor model checkpoint") actor_booster.save_model(actor, os.path.join(trainer.actor_save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final actor model checkpoint at episodes {args.num_episodes} at folder {args.save_path}" ) coordinator.print_on_master("Start saving final critic model checkpoint") critic_booster.save_model(critic, os.path.join(trainer.critic_save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final critic model checkpoint at episodes {args.num_episodes} at folder {args.save_path}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_dataset", nargs="+", default=[]) parser.add_argument("--ptx_dataset", nargs="+", default=[]) parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d"], help="Choose which plugin to use", ) parser.add_argument( "--conversation_template_config", type=str, default=None, help="Path \ to save conversation template config files.", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--rm_pretrain", type=str, default=None) parser.add_argument("--no_neural_reward_model", default=False, action="store_true") parser.add_argument("--checkpoint_path", type=str, default=None) parser.add_argument("--critic_checkpoint_path", type=str, default=None) parser.add_argument("--rm_checkpoint_path", type=str, help="Reward model checkpoint path") parser.add_argument("--reward_functions", type=str, nargs="+", default=None, help="Reward functions to use") parser.add_argument("--save_path", type=str, default="actor_checkpoint_prompts") parser.add_argument("--num_episodes", type=int, default=1) parser.add_argument("--num_collect_steps", type=int, default=2) parser.add_argument("--num_update_steps", type=int, default=5) parser.add_argument("--save_interval", type=int, default=1000) parser.add_argument("--train_batch_size", type=int, default=16) parser.add_argument("--experience_batch_size", type=int, default=16) parser.add_argument("--ptx_batch_size", type=int, default=4) parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--lr", type=float, default=9e-6) parser.add_argument("--critic_lr", type=float, default=9e-6) parser.add_argument("--kl_coef", type=float, default=0.1) parser.add_argument("--ptx_coef", type=float, default=0.0) parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--max_length", type=int, default=2048) parser.add_argument("--max_seq_len", type=int, default=256) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_orpo.py
applications/ColossalChat/examples/training_scripts/train_orpo.py
import argparse import json import os import resource from contextlib import nullcontext import torch from coati.dataset import DataCollatorForPreferenceDataset, StatefulDistributedSampler, load_tokenized_dataset from coati.models import LoraConfig, convert_to_lora_module, disable_dropout from coati.trainer import ORPOTrainer from coati.utils import load_checkpoint from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam logger = get_dist_logger() def train(args): lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=True) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: model = AutoModelForCausalLM.from_pretrained(args.pretrain) if args.lora_config is not None: model = convert_to_lora_module(model, lora_config=lora_config) for name, module in model.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) disable_dropout(model) if args.grad_checkpoint: # Note, for some models, lora may not be compatible with gradient checkpointing model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False # configure optimizer optim = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # configure dataset coordinator.print_on_master(f"Load dataset: {args.dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_dataset = load_tokenized_dataset(dataset_paths=args.dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) train_dataloader = plugin.prepare_dataloader( dataset=train_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) eval_dataloader = None if args.eval_dataset: eval_dataset = load_tokenized_dataset(dataset_paths=args.eval_dataset, mode="dev") eval_data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) eval_dataloader = plugin.prepare_dataloader( dataset=eval_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=eval_data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: logger.warning("No evaluation dataset is provided, skip evaluation") num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps if args.warmup_steps is None: args.warmup_steps = int(args.max_epochs * 0.025 * (len(train_dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optim, total_steps=args.max_epochs * num_update_steps_per_epoch, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optim, _, train_dataloader, lr_scheduler = booster.boost( model=model, optimizer=optim, lr_scheduler=lr_scheduler, dataloader=train_dataloader, ) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 sampler_start_idx = 0 start_step = 0 if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: coordinator.print_on_master(f"Continued pretrain from checkpoint {args.checkpoint_path}") booster.load_model(model, args.checkpoint_path) else: coordinator.print_on_master(f"Load model checkpoint from {args.checkpoint_path}") start_epoch, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=booster, model=model, optimizer=optim, lr_scheduler=lr_scheduler, ) assert isinstance(train_dataloader.sampler, StatefulDistributedSampler) train_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded checkpoint {args.checkpoint_path} at epoch {start_epoch} step {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) trainer = ORPOTrainer( actor=model, booster=booster, actor_optim=optim, plugin=plugin, actor_lr_scheduler=lr_scheduler, tokenizer=tokenizer, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, start_epoch=start_epoch, save_interval=args.save_interval, save_dir=args.save_dir, coordinator=coordinator, lam=args.lam, apply_loss_mask=not args.disable_loss_mask, ) trainer.fit( train_preference_dataloader=train_dataloader, eval_preference_dataloader=eval_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights model.eval() # save model checkpoint after fitting on only rank0 if args.save_dir is not None: coordinator.print_on_master("Start saving final model checkpoint") booster.save_model(model, os.path.join(args.save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final model checkpoint at epoch {args.max_epochs} at folder {args.save_dir}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d"], help="Choose which plugin to use", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--lam", type=float, default=0.1, help="lambda in ORPO loss") parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--model_type", type=str, default=None) parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--dataset", nargs="+", default=[]) parser.add_argument("--eval_dataset", nargs="+", default=[]) parser.add_argument( "--checkpoint_path", type=str, default=None, help="Checkpoint path if need to resume training form a checkpoint" ) parser.add_argument("--config_file", type=str, default=None, help="Config file") parser.add_argument("--save_dir", type=str, default=None) parser.add_argument("--max_length", type=int, default=2048, help="Model max length") parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument( "--disable_reference_model", action="store_true", default=False, help="Disable the reference model (enabled by default)", ) parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--save_interval", type=int, default=1000, help="number of step between two checkpoints") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() if args.config_file is not None: os.makedirs(os.path.dirname(args.config_file), exist_ok=True) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_kto.py
applications/ColossalChat/examples/training_scripts/train_kto.py
import argparse import json import os import resource from contextlib import nullcontext import torch from coati.dataset import DataCollatorForKTODataset, StatefulDistributedSampler, load_tokenized_dataset from coati.models import LoraConfig, convert_to_lora_module, disable_dropout from coati.trainer import KTOTrainer from coati.utils import load_checkpoint from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam logger = get_dist_logger() def train(args): lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=True) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) ref_booster = Booster(plugin=plugin) # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: model = AutoModelForCausalLM.from_pretrained(args.pretrain) if args.use_flash_attn: ref_model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) else: ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain) if args.lora_config is not None: model = convert_to_lora_module(model, lora_config=lora_config) for name, module in model.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) disable_dropout(ref_model) disable_dropout(model) if args.grad_checkpoint: # Note, for some models, lora may not be compatible with gradient checkpointing model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False # configure optimizer optim = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # configure dataset coordinator.print_on_master(f"Load dataset: {args.dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_dataset = load_tokenized_dataset(dataset_paths=args.dataset, mode="train", mode_map=mode_map) num_desirable = 0 num_undesirable = 0 for i in range(len(train_dataset)): if train_dataset[i]["label"]: num_desirable += 1 else: num_undesirable += 1 logger.info(f"Dataset Statistics:\nDesirable: {num_desirable}\nUndesirable: {num_undesirable}") # Check if the user specified weights fit into the theoratical lower and upper bounds from Eq. (8) of https://arxiv.org/abs/2402.01306 actual_ratio = (args.desirable_weight * num_desirable) / (args.undesirable_weight * num_undesirable) if actual_ratio < 1 or actual_ratio > 4 / 3: if not args.auto_weight: raise AssertionError( f"Desirable weight and undesirable weight are not within the theoratical bounds, [1, 4/3]. Actual ratio: {actual_ratio}, please increase/decrease desirable weight or decrease/increase undesirable weight." ) else: args.desirable_weight = args.desirable_weight / actual_ratio coordinator.print_on_master( f"Desirable weight and undesirable weight are not within the theoratical bounds, [1, 4/3]. Actual ratio: {actual_ratio}, auto weight is enabled, set desirable weight to {args.desirable_weight} and undesirable weight to {args.undesirable_weight}" ) data_collator = DataCollatorForKTODataset(tokenizer=tokenizer, max_length=args.max_length) train_dataloader = plugin.prepare_dataloader( dataset=train_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) eval_dataloader = None if args.eval_dataset: eval_dataset = load_tokenized_dataset(dataset_paths=args.eval_dataset, mode="dev") eval_data_collator = DataCollatorForKTODataset(tokenizer=tokenizer, max_length=args.max_length) eval_dataloader = plugin.prepare_dataloader( dataset=eval_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=eval_data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: logger.warning("No evaluation dataset is provided, skip evaluation") num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps if args.warmup_steps is None: args.warmup_steps = int(args.max_epochs * 0.025 * (len(train_dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optim, total_steps=args.max_epochs * num_update_steps_per_epoch, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optim, _, train_dataloader, lr_scheduler = booster.boost( model=model, optimizer=optim, lr_scheduler=lr_scheduler, dataloader=train_dataloader, ) if ref_model is not None: ref_model, _, _, _, _ = ref_booster.boost(model=ref_model, dataloader=train_dataloader) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 sampler_start_idx = 0 start_step = 0 if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: coordinator.print_on_master(f"Continued pretrain from checkpoint {args.checkpoint_path}") booster.load_model(model, args.checkpoint_path) else: coordinator.print_on_master(f"Load model checkpoint from {args.checkpoint_path}") start_epoch, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=booster, model=model, optimizer=optim, lr_scheduler=lr_scheduler, ) assert isinstance(train_dataloader.sampler, StatefulDistributedSampler) train_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded checkpoint {args.checkpoint_path} at epoch {start_epoch} step {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) trainer = KTOTrainer( actor=model, ref_model=ref_model, booster=booster, actor_optim=optim, plugin=plugin, actor_lr_scheduler=lr_scheduler, tokenizer=tokenizer, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, start_epoch=start_epoch, save_interval=args.save_interval, save_dir=args.save_dir, coordinator=coordinator, beta=args.beta, desirable_weight=args.desirable_weight, undesirable_weight=args.undesirable_weight, apply_loss_mask=not args.disable_loss_mask, ) trainer.fit( train_preference_dataloader=train_dataloader, eval_preference_dataloader=eval_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights model.eval() # save model checkpoint after fitting on only rank0 if args.save_dir is not None: coordinator.print_on_master("Start saving final model checkpoint") booster.save_model(model, os.path.join(args.save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final model checkpoint at epoch {args.max_epochs} at folder {args.save_dir}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d"], help="Choose which plugin to use", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--beta", type=float, default=0.1, help="beta in KTO loss") parser.add_argument("--desirable_weight", type=float, default=1.0, help="desirable_weight in KTO loss") parser.add_argument("--undesirable_weight", type=float, default=1.0, help="undesirable_weight in KTO loss") parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--dataset", nargs="+", default=[]) parser.add_argument("--eval_dataset", nargs="+", default=[]) parser.add_argument( "--checkpoint_path", type=str, default=None, help="Checkpoint path if need to resume training form a checkpoint" ) parser.add_argument("--config_file", type=str, default=None, help="Config file") parser.add_argument("--save_dir", type=str, default=None) parser.add_argument("--max_length", type=int, default=2048, help="Model max length") parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--save_interval", type=int, default=1000, help="number of step between two checkpoints") parser.add_argument("--auto_weight", default=False, action="store_true") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() if args.config_file is not None: os.makedirs(os.path.dirname(args.config_file), exist_ok=True) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_sft.py
applications/ColossalChat/examples/training_scripts/train_sft.py
import argparse import json import math import os import resource from contextlib import nullcontext import torch from coati.dataset import DataCollatorForSupervisedDataset, StatefulDistributedSampler, load_tokenized_dataset from coati.models import LoraConfig, convert_to_lora_module from coati.trainer import SFTTrainer from coati.utils import load_checkpoint from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam logger = get_dist_logger() def train(args): lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ============================== # Initialize Booster # ============================== init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, attn_implementation="flash_attention_2", trust_remote_code=True, ) else: model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, trust_remote_code=True, ) if lora_config is not None: model = convert_to_lora_module(model, lora_config=lora_config) for name, module in model.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=not args.grad_checkpoint) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True if args.accumulation_steps > 1 else False, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, microbatch_size=args.microbatch_size, ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) # configure optimizer optim = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) if args.grad_checkpoint: # Note, for some models, lora may not be compatible with gradient checkpointing model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer = AutoTokenizer.from_pretrained( args.tokenizer_dir or args.pretrain, use_fast=False, trust_remote_code=True ) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False tokenizer.padding_side = "right" coordinator.print_on_master(f"Configuration file will be saved at: {args.config_file}") coordinator.print_on_master(f"Model checkpoint will be saved at: {args.save_path}") # configure dataset coordinator.print_on_master( f"Max CUDA memory before data loader: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) dataset = load_tokenized_dataset(dataset_paths=args.dataset, mode="train") data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer, max_length=args.max_len) train_dataloader = plugin.prepare_dataloader( dataset=dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) eval_dataloader = None if args.eval_dataset: eval_dataset = load_tokenized_dataset(dataset_paths=args.eval_dataset, mode="dev") eval_data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer, max_length=args.max_len) eval_dataloader = plugin.prepare_dataloader( dataset=eval_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=eval_data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: logger.warning("No evaluation dataset is provided, skip evaluation") coordinator.print_on_master( f"Max CUDA memory after data loader: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps math.ceil(args.max_epochs * num_update_steps_per_epoch) if args.warmup_steps is None: args.warmup_steps = int(args.max_epochs * 0.025 * (len(train_dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optim, total_steps=args.max_epochs * num_update_steps_per_epoch, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) # Flash attention will be disabled because it does NOT support fp32. default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optim, _, train_dataloader, lr_scheduler = booster.boost( model=model, optimizer=optim, lr_scheduler=lr_scheduler, dataloader=train_dataloader, ) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 sampler_start_idx = 0 start_step = 0 if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: coordinator.print_on_master(f"Continued pretrain from checkpoint {args.checkpoint_path}") booster.load_model(model, args.checkpoint_path) else: coordinator.print_on_master(f"Load model checkpoint from {args.checkpoint_path}") start_epoch, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=booster, model=model, optimizer=optim, lr_scheduler=lr_scheduler, ) train_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded checkpoint {args.checkpoint_path} at epoch {start_epoch} step {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) trainer = SFTTrainer( model=model, booster=booster, optim=optim, plugin=plugin, lr_scheduler=lr_scheduler, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, apply_loss_mask=not args.disable_loss_mask, start_epoch=start_epoch, save_interval=args.save_interval, save_dir=args.save_path, coordinator=coordinator, ) trainer.fit( train_dataloader=train_dataloader, eval_dataloader=eval_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights model.eval() # save model checkpoint after fitting on only rank0 if args.save_path is not None: coordinator.print_on_master("Start saving final model checkpoint") booster.save_model(model, os.path.join(args.save_path, "modeling"), shard=True) coordinator.print_on_master( f"Saved final model checkpoint at epoch {args.max_epochs} at folder {args.save_path}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "3d", "ddp", "zero2_cpu", "zero2"], help="Choose which plugin to use", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--dataset", nargs="+", default=[]) parser.add_argument("--eval_dataset", nargs="+", default=[]) parser.add_argument( "--checkpoint_path", type=str, default=None, help="Checkpoint path if need to resume training form a checkpoint" ) parser.add_argument("--save_path", type=str, default=None) parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--max_len", type=int, default=512) parser.add_argument("--mixed_precision", type=str, default="bf16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--save_interval", type=int, default=1000, help="number of step between two checkpoints") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--config_file", type=str, default=None, help="Config file") parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") parser.add_argument("--microbatch_size", type=int, default=1) args = parser.parse_args() if args.config_file is not None: os.makedirs(os.path.dirname(args.config_file), exist_ok=True) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_grpo.py
applications/ColossalChat/examples/training_scripts/train_grpo.py
import argparse import json import os import resource from contextlib import nullcontext import torch import torch.distributed as dist from coati.dataset import ( DataCollatorForPromptDataset, DataCollatorForSupervisedDataset, StatefulDistributedSampler, load_tokenized_dataset, setup_conversation_template, ) from coati.models import LoraConfig, RewardModel, RLVRRewardModel, convert_to_lora_module, disable_dropout, lora_manager from coati.trainer import GRPOTrainer from coati.utils import load_checkpoint from coati.utils.reward_score import * from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam from colossalai.shardformer.policies.auto_policy import get_autopolicy logger = get_dist_logger() # default settings for response format tags, overwrite it in chat_template definition if needed response_format_tags = { "think_start": {"text": "<think>", "num_occur": 1}, "think_end": {"text": "</think>", "num_occur": 1}, "answer_start": {"text": "<answer>", "num_occur": 1}, "answer_end": {"text": "</answer>", "num_occur": 1}, } def train(args): global response_format_tags lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: actor = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) ref_model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) if args.rm_pretrain: reward_model = RewardModel( args.rm_pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, trust_remote_code=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: actor = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True) if args.rm_pretrain: reward_model = RewardModel(args.rm_pretrain, trust_remote_code=True) ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain, trust_remote_code=True) if args.lora_config is not None: actor = convert_to_lora_module(actor, lora_config=lora_config) for name, module in actor.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) lora_manager.able_to_merge = False # Disable dropout disable_dropout(actor) if args.grad_checkpoint: actor.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if os.path.exists(args.conversation_template_config): with open(args.conversation_template_config, "r", encoding="utf8") as f: conversation_template_config = json.load(f) dist.barrier() if "response_format_tags" in conversation_template_config: logger.warning(f"Overwrite default response format tags with {args.conversation_template_config}") response_format_tags = conversation_template_config.get("response_format_tags", response_format_tags) conversation_template = setup_conversation_template( tokenizer, chat_template_config=conversation_template_config, save_path=args.conversation_template_config ) stop_ids = conversation_template.stop_ids if len(conversation_template.stop_ids) > 0 else None else: raise ValueError("Conversation template config is not provided or incorrect") if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False tokenizer.padding_side = "left" # left padding for generation (online learning) # configure generation config actor.generation_config.update( pad_token_id=tokenizer.eos_token_id, bos_token_id=tokenizer.bos_token_id, eos_token_id=tokenizer.eos_token_id ) # configure optimizer coordinator.print_on_master(f"setting up optimizer for actor: lr={args.lr}, weight_decay={args.weight_decay}") actor_optim = HybridAdam( model_params=actor.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) if args.warmup_steps is None: args.warmup_steps = int(0.025 * args.num_episodes) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") actor_lr_scheduler = CosineAnnealingWarmupLR( optimizer=actor_optim, total_steps=args.num_episodes, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=True) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": if args.use_flash_attn and (args.tp > 1 or args.pp > 1 or args.sp > 1 or args.enable_sequence_parallelism): logger.warning("Flash attention cannot be used with 3D parallelism for PPO training. Disabling it.") args.use_flash_attn = False plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, ) if args.rm_pretrain: custom_plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, custom_policy=get_autopolicy(reward_model.model), ) else: raise ValueError(f"Unknown plugin {args.plugin}") if args.plugin != "3d" and args.rm_pretrain: custom_plugin = plugin # configure dataset coordinator.print_on_master(f"Load dataset: {args.prompt_dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_prompt_dataset = load_tokenized_dataset(dataset_paths=args.prompt_dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForPromptDataset(tokenizer=tokenizer, max_length=args.max_length - args.max_seq_len) train_prompt_dataloader = plugin.prepare_dataloader( dataset=train_prompt_dataset, batch_size=args.experience_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) if len(args.ptx_dataset) > 0: train_ptx_dataset = load_tokenized_dataset(dataset_paths=args.ptx_dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForSupervisedDataset(tokenizer=tokenizer, max_length=args.max_length) train_pretrain_dataloader = plugin.prepare_dataloader( dataset=train_ptx_dataset, batch_size=args.ptx_batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: train_pretrain_dataloader = None actor_booster = Booster(plugin=plugin) ref_booster = Booster(plugin=plugin) if args.rm_pretrain: rm_booster = Booster(plugin=custom_plugin) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) actor, actor_optim, _, train_prompt_dataloader, actor_lr_scheduler = actor_booster.boost( model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, dataloader=train_prompt_dataloader, ) if args.rm_pretrain: reward_model, _, _, _, _ = rm_booster.boost(model=reward_model, dataloader=train_prompt_dataloader) else: if args.reward_functions: reward_fn_list = [] for reward_fn in args.reward_functions: """ To define custom reward function, you can define your functions under: colossalai/applications/ColossalChat/coati/utils/reward_score/__init__.py and use it here by mofiying the following line: """ if reward_fn == "gsm8k_reward_fn": reward_fn_list.append(gsm8k_reward_fn) elif reward_fn == "math_competition_reward_fn": reward_fn_list.append(math_competition_reward_fn) else: raise ValueError(f"Unknown reward function {reward_fn}") reward_model = RLVRRewardModel( reward_fn_list=reward_fn_list, tokenizer=tokenizer, tags=response_format_tags ) ref_model, _, _, _, _ = ref_booster.boost(model=ref_model, dataloader=train_prompt_dataloader) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) sampler_start_idx = 0 start_step = 0 if args.rm_checkpoint_path is not None: if "modeling" in args.rm_checkpoint_path: rm_booster.load_model(reward_model, args.rm_checkpoint_path) else: _, _, _ = load_checkpoint( load_dir=args.rm_checkpoint_path, booster=rm_booster, model=reward_model, optimizer=None, lr_scheduler=None, ) coordinator.print_on_master(f"Loaded reward model checkpoint {args.rm_checkpoint_path}") if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: actor_booster.load_model(actor, args.checkpoint_path) ref_booster.load_model(ref_model, args.checkpoint_path) coordinator.print_on_master(f"Loaded actor and reference model {args.checkpoint_path}") else: _, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=actor_booster, model=actor, optimizer=actor_optim, lr_scheduler=actor_lr_scheduler, ) _, _, _ = load_checkpoint(load_dir=args.checkpoint_path, booster=ref_booster, model=ref_model) assert isinstance(train_prompt_dataloader.sampler, StatefulDistributedSampler) train_prompt_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded actor and reference model checkpoint {args.checkpoint_path} at spisode {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) # configure trainer trainer = GRPOTrainer( actor_booster, actor, reward_model, ref_model, actor_optim, actor_lr_scheduler, tokenizer=tokenizer, stop_token_ids=[stop_ids], kl_coef=args.kl_coef, ptx_coef=args.ptx_coef, train_batch_size=args.train_batch_size, buffer_limit=args.num_collect_steps * args.experience_batch_size * args.num_generations, max_length=args.max_length, use_cache=True, do_sample=True, apply_loss_mask=not args.disable_loss_mask, accumulation_steps=args.accumulation_steps, save_dir=args.save_path, save_interval=args.save_interval, top_k=50, use_tp=args.tp > 1, num_generations=args.num_generations, inference_batch_size=args.inference_batch_size, logits_forward_batch_size=args.logits_forward_batch_size, offload_inference_models="gemini" not in args.plugin, coordinator=coordinator, max_tokens_thinking=args.max_tokens_thinking if args.max_tokens_thinking else args.max_length - 100, temperature_annealing_config={ "start_temperature": args.initial_temperature, "end_temperature": args.final_temperature, "annealing_warmup_steps": min(100, int(args.num_episodes / 6)), "annealing_steps": min(600, int(args.num_episodes / 2)), }, # Hack: some old model's default update_model_kwargs_fn/prepare_inputs_fn may doesn't work due to version conflict with transformers, you can overwrite them # update_model_kwargs_fn=update_model_kwargs_fn, # prepare_inputs_fn = None ) trainer.fit( num_episodes=args.num_episodes, num_collect_steps=args.num_collect_steps, num_update_steps=args.num_update_steps, prompt_dataloader=train_prompt_dataloader, pretrain_dataloader=train_pretrain_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights lora_manager.able_to_merge = True actor.eval() # save model checkpoint after fitting on only rank0 coordinator.print_on_master("Start saving final actor model checkpoint") actor_booster.save_model(actor, os.path.join(trainer.actor_save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final actor model checkpoint at episodes {args.num_episodes} at folder {args.save_path}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_dataset", nargs="+", default=[]) parser.add_argument("--ptx_dataset", nargs="+", default=[]) parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d"], help="Choose which plugin to use", ) parser.add_argument( "--conversation_template_config", type=str, default=None, help="Path \ to save conversation template config files.", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--rm_pretrain", type=str, default=None) parser.add_argument("--checkpoint_path", type=str, default=None) parser.add_argument("--rm_checkpoint_path", type=str, help="Reward model checkpoint path") parser.add_argument("--reward_functions", type=str, nargs="+", default=None, help="Reward functions to use") parser.add_argument("--save_path", type=str, default="actor_checkpoint_prompts") parser.add_argument("--num_episodes", type=int, default=1) parser.add_argument("--num_collect_steps", type=int, default=2) parser.add_argument("--num_update_steps", type=int, default=5) parser.add_argument("--num_generations", type=int, default=8) parser.add_argument("--inference_batch_size", type=int, default=None) parser.add_argument("--save_interval", type=int, default=1000) parser.add_argument("--train_batch_size", type=int, default=16) parser.add_argument("--logits_forward_batch_size", type=int, default=1) parser.add_argument("--experience_batch_size", type=int, default=16) parser.add_argument("--ptx_batch_size", type=int, default=4) parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--lr", type=float, default=1e-6) parser.add_argument("--kl_coef", type=float, default=0.7) parser.add_argument("--ptx_coef", type=float, default=0.0) parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--max_length", type=int, default=2048) parser.add_argument("--max_tokens_thinking", type=int, default=2000) parser.add_argument("--max_seq_len", type=int, default=256) parser.add_argument("--initial_temperature", type=float, default=1.0) parser.add_argument("--final_temperature", type=float, default=0.9) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_dpo.py
applications/ColossalChat/examples/training_scripts/train_dpo.py
import argparse import json import os import resource from contextlib import nullcontext import torch from coati.dataset import DataCollatorForPreferenceDataset, StatefulDistributedSampler, load_tokenized_dataset from coati.models import LoraConfig, convert_to_lora_module, disable_dropout from coati.trainer import DPOTrainer from coati.utils import load_checkpoint from transformers import AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam logger = get_dist_logger() def train(args): lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=not args.grad_checkpoint) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=True, enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, microbatch_size=args.microbatch_size, ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) ref_plugin = HybridParallelPlugin( tp_size=args.ref_tp, pp_size=1, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, ) ref_booster = Booster(plugin=ref_plugin) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: model = AutoModelForCausalLM.from_pretrained(args.pretrain) if not args.disable_reference_model: if args.use_flash_attn: ref_model = AutoModelForCausalLM.from_pretrained( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) else: ref_model = AutoModelForCausalLM.from_pretrained(args.pretrain) else: ref_model = None if args.lora_config is not None: model = convert_to_lora_module(model, lora_config=lora_config) for name, module in model.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) disable_dropout(model) disable_dropout(ref_model) if args.grad_checkpoint: # Make sure gradient checkpointing can be activated. model.train() # Note, for some models, lora may not be compatible with gradient checkpointing. model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.add_bos_token = False tokenizer.add_eos_token = False # configure optimizer optim = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # Configure dataset coordinator.print_on_master(f"Load dataset: {args.dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_dataset = load_tokenized_dataset(dataset_paths=args.dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) train_dataloader = plugin.prepare_dataloader( dataset=train_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) eval_dataloader = None if args.eval_dataset: eval_dataset = load_tokenized_dataset(dataset_paths=args.eval_dataset, mode="dev") eval_data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) eval_dataloader = plugin.prepare_dataloader( dataset=eval_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=eval_data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: logger.warning("No evaluation dataset is provided, skip evaluation") num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps if args.warmup_steps is None: args.warmup_steps = int(args.max_epochs * 0.025 * (len(train_dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optim, total_steps=args.max_epochs * num_update_steps_per_epoch, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optim, _, train_dataloader, lr_scheduler = booster.boost( model=model, optimizer=optim, lr_scheduler=lr_scheduler, dataloader=train_dataloader, ) ref_model, _, _, _, _ = ref_booster.boost(model=ref_model) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 sampler_start_idx = 0 start_step = 0 if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: coordinator.print_on_master(f"Continued pretrain from checkpoint {args.checkpoint_path}") booster.load_model(model, args.checkpoint_path) else: coordinator.print_on_master(f"Load model checkpoint from {args.checkpoint_path}") start_epoch, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=booster, model=model, optimizer=optim, lr_scheduler=lr_scheduler, ) assert isinstance(train_dataloader.sampler, StatefulDistributedSampler) train_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded checkpoint {args.checkpoint_path} at epoch {start_epoch} step {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) trainer = DPOTrainer( actor=model, ref_model=ref_model, booster=booster, actor_optim=optim, plugin=plugin, actor_lr_scheduler=lr_scheduler, tokenizer=tokenizer, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, start_epoch=start_epoch, save_interval=args.save_interval, save_dir=args.save_dir, coordinator=coordinator, beta=args.beta, gamma=args.gamma, length_normalization=args.length_normalization, apply_loss_mask=not args.disable_loss_mask, ) trainer.fit( train_preference_dataloader=train_dataloader, eval_preference_dataloader=eval_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights model.eval() # save model checkpoint after fitting on only rank0 if args.save_dir is not None: coordinator.print_on_master("Start saving final model checkpoint") booster.save_model(model, os.path.join(args.save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final model checkpoint at epoch {args.max_epochs} at folder {args.save_dir}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "zero2", "zero2_cpu", "3d", "ddp"], help="Choose which plugin to use", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--loss_type", type=str, default="dpo_loss", help="dpo_loss or simpo_loss") parser.add_argument("--beta", type=float, default=0.1, help="beta in DPO loss") parser.add_argument("--gamma", type=float, default=0.0, help="gamma in SimPO loss") parser.add_argument("--length_normalization", default=False, action="store_true") parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--model_type", type=str, default=None) parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--dataset", nargs="+", default=[]) parser.add_argument("--eval_dataset", nargs="+", default=[]) parser.add_argument( "--checkpoint_path", type=str, default=None, help="Checkpoint path if need to resume training form a checkpoint" ) parser.add_argument("--config_file", type=str, default=None, help="Config file") parser.add_argument("--save_dir", type=str, default=None) parser.add_argument("--max_length", type=int, default=2048, help="Model max length") parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--disable_loss_mask", default=False, action="store_true") parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--save_interval", type=int, default=1000, help="number of step between two checkpoints") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--accumulation_steps", type=int, default=1) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") parser.add_argument( "--microbatch_size", type=int, default=2, help="Micro batch size for PP training. To activate PP training for DPO-like algorithm, you must keep size even and the size should be equal or greater than 2.", ) # Parameter for reference model parser.add_argument( "--disable_reference_model", action="store_true", default=False, help="Disable the reference model (enabled by default)", ) parser.add_argument( "--ref_tp", type=int, default=1, help="TP size for reference model; used only when reference model is too large.", ) args = parser.parse_args() # fool proof hyperparameter setup if args.loss_type == "simpo_loss": args.length_normalization = True args.gamma = args.gamma if args.gamma > 0 else 1.4 if args.config_file is not None: os.makedirs(os.path.dirname(args.config_file), exist_ok=True) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/lora_finetune.py
applications/ColossalChat/examples/training_scripts/lora_finetune.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Supervised fine-tuning of MoE models like Deepseek V3/R1 on a downstream task. """ import argparse import json import os import resource from contextlib import nullcontext from types import MethodType import torch import torch.distributed as dist from coati.dataset.loader import RawConversationDataset from peft import LoraConfig from tqdm import tqdm from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossalai.booster.plugin import ( GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin, MoeHybridParallelPlugin, Plugin, TorchDDPPlugin, ) from colossalai.cluster import DistCoordinator from colossalai.lazy import LazyInitContext from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam from colossalai.utils import get_current_device def all_reduce_mean(loss: torch.Tensor, plugin: Plugin) -> torch.Tensor: loss = loss.data group = getattr(plugin, "dp_group", None) dist.all_reduce(loss, group=group) return loss / dist.get_world_size(group) def train(args) -> None: # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() accelerator = get_accelerator() coordinator = DistCoordinator() # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": plugin = TorchDDPPlugin(find_unused_parameters=True if args.use_grad_checkpoint is False else False) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=(args.accumulation_steps > 1), enable_fused_normalization=get_accelerator().is_available(), enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, max_norm=args.grad_clip, enable_gradient_accumulation=(args.accumulation_steps > 1), enable_fused_normalization=get_accelerator().is_available(), enable_flash_attention=args.use_flash_attn, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_fused_normalization=get_accelerator().is_available(), enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, max_norm=args.grad_clip, precision=args.mixed_precision, microbatch_size=args.microbatch_size, ) elif args.plugin == "moe": plugin = MoeHybridParallelPlugin( ep_size=args.ep, tp_size=args.tp, pp_size=args.pp, zero_stage=args.zero_stage, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, enable_sequence_parallelism=args.sp > 1, enable_fused_normalization=get_accelerator().is_available(), enable_flash_attention=args.use_flash_attn, max_norm=args.grad_clip, precision=args.mixed_precision, microbatch_size=args.microbatch_size, ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) def is_master(): if isinstance(plugin, HybridParallelPlugin) and plugin.pp_size > 1: return coordinator.rank == coordinator.world_size - 1 return coordinator.is_master() # ============================== # Initialize Tensorboard and Save Config # ============================== if is_master(): if args.tensorboard_dir is not None: from torch.utils.tensorboard import SummaryWriter os.makedirs(args.tensorboard_dir, exist_ok=True) writer = SummaryWriter(args.tensorboard_dir) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) # ====================================================== # Initialize Tokenizer, Dataset, Collator and Dataloader # ====================================================== tokenizer = AutoTokenizer.from_pretrained(args.pretrained, trust_remote_code=True) coordinator.print_on_master( f"Training Info:\nConfig file: {args.config_file} \nTensorboard logs: {args.tensorboard_dir} \nModel checkpoint: {args.save_dir}" ) coordinator.print_on_master(f"Load dataset: {args.dataset}") dataset = RawConversationDataset( tokenizer, args.dataset, args.max_length, ) dataloader = plugin.prepare_dataloader( dataset=dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, ) coordinator.print_on_master( f"Max device memory after data loader: {accelerator.max_memory_allocated() / 1024 ** 2:.2f} MB" ) # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # When training the ChatGLM model, LoRA and gradient checkpointing are incompatible. init_ctx = ( LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin, HybridParallelPlugin)) else nullcontext() ) attn_impl = "eager" if get_accelerator().name == "npu" else "flash_attention_2" config = AutoConfig.from_pretrained(args.pretrained, trust_remote_code=True) with init_ctx: # from_pretrained is not compatible with LoRA, we load pretrained weights later. # model = AutoModelForCausalLM.from_pretrained( # args.pretrained, # torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, # trust_remote_code=True, # attn_implementation=attn_impl, # ) model = AutoModelForCausalLM.from_config( config, trust_remote_code=True, attn_implementation=attn_impl, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, ) if args.lora_rank > 0: if model.__class__.__name__.startswith("DeepseekV3"): lora_config = LoraConfig( task_type="CAUSAL_LM", r=args.lora_rank, lora_alpha=args.lora_alpha, target_modules=["gate_proj", "up_proj", "down_proj"], ) else: lora_config = LoraConfig(task_type="CAUSAL_LM", r=args.lora_rank, lora_alpha=args.lora_alpha) model = booster.enable_lora(model, lora_config=lora_config) # this is essential, otherwise the grad checkpoint will not work. model.train() if args.use_grad_checkpoint: model.gradient_checkpointing_enable() coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") if model.config.__class__.__name__.startswith("DeepseekV3"): model.config.use_cache = False model.eval() # enable grad for moe layers for m in model.modules(): if m.__class__.__name__ == "DeepseekV3MoE": m.moe_infer = MethodType(m.moe_infer.__wrapped__, m) model_numel = sum(p.numel() for p in model.parameters()) coordinator.print_on_master(f"Model params: {model_numel / 1e9:.2f} B") optimizer = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) if args.warmup_steps is None: args.warmup_steps = int(args.num_epochs * 0.025 * (len(dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optimizer, total_steps=args.num_epochs * (len(dataloader) // args.accumulation_steps), warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) # Flash attention will be disabled because it does NOT support fp32. default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optimizer, _, dataloader, lr_scheduler = booster.boost( model=model, optimizer=optimizer, lr_scheduler=lr_scheduler, dataloader=dataloader, ) torch.set_default_dtype(torch.float) booster.load_model(model, args.pretrained, low_cpu_mem_mode=False, num_threads=8) coordinator.print_on_master( f"Booster init max device memory: {accelerator.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 start_step = 0 num_steps_per_epoch = len(dataloader) // args.accumulation_steps for epoch in range(start_epoch, args.num_epochs): dataloader.sampler.set_epoch(epoch=epoch) if isinstance(plugin, HybridParallelPlugin) and plugin.pp_size > 1: data_iter = iter(dataloader) step_bar = tqdm( range(len(dataloader)), desc="Step", disable=not is_master(), ) for step in step_bar: outputs = booster.execute_pipeline( data_iter, model, criterion=lambda outputs, inputs: outputs[0], optimizer=optimizer, return_loss=True, ) loss = outputs["loss"] if booster.plugin.stage_manager.is_last_stage(): global_loss = all_reduce_mean(loss, plugin) optimizer.step() if booster.plugin.stage_manager.is_last_stage(): grad_norm = optimizer.get_grad_norm() step_bar.set_postfix({"loss": global_loss.item(), "grad_norm": grad_norm}) if args.tensorboard_dir is not None and is_master(): global_step = (epoch * num_steps_per_epoch) + (step + 1) // args.accumulation_steps writer.add_scalar(tag="Loss", scalar_value=global_loss.item(), global_step=global_step) writer.add_scalar( tag="Learning Rate", scalar_value=lr_scheduler.get_last_lr()[0], global_step=global_step, ) writer.add_scalar(tag="Grad Norm", scalar_value=grad_norm, global_step=global_step) lr_scheduler.step() optimizer.zero_grad() else: pbar = tqdm( dataloader, desc=f"Epoch {epoch}", disable=not is_master(), initial=start_step // args.accumulation_steps, ) total_loss = torch.tensor(0.0, device=get_current_device()) for step, batch in enumerate(pbar, start=start_step // args.accumulation_steps): batch = {k: v.to(get_current_device()) for k, v in batch.items() if isinstance(v, torch.Tensor)} batch_output = model(**batch) loss = batch_output.loss / args.accumulation_steps total_loss.add_(loss.data) booster.backward(loss=loss, optimizer=optimizer) if (step + 1) % args.accumulation_steps == 0: all_reduce_mean(total_loss, plugin) optimizer.step() grad_norm = optimizer.get_grad_norm() pbar.set_postfix({"loss": total_loss.item(), "grad_norm": grad_norm}) if args.tensorboard_dir is not None and is_master(): global_step = (epoch * num_steps_per_epoch) + (step + 1) // args.accumulation_steps writer.add_scalar(tag="Loss", scalar_value=total_loss.item(), global_step=global_step) writer.add_scalar( tag="Learning Rate", scalar_value=lr_scheduler.get_last_lr()[0], global_step=global_step, ) writer.add_scalar(tag="Grad Norm", scalar_value=grad_norm, global_step=global_step) lr_scheduler.step() optimizer.zero_grad() total_loss.fill_(0.0) # Delete cache. # del batch, batch_labels, batch_output, loss accelerator.empty_cache() # Final save. coordinator.print_on_master("Start saving final model checkpoint") if args.lora_rank > 0: booster.save_lora_as_pretrained(model, os.path.join(args.save_dir, "lora")) else: booster.save_model(model, os.path.join(args.save_dir, "modeling"), shard=True) coordinator.print_on_master(f"Saved final model checkpoint at epoch {epoch} at folder {args.save_dir}") coordinator.print_on_master(f"Max device memory usage: {accelerator.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": parser = argparse.ArgumentParser() # Basic training information. parser.add_argument( "-m", "--pretrained", type=str, required=True, help="Address of the pre-trained model", ) parser.add_argument("-d", "--dataset", type=str, required=True, help="Raw Jonl dataset for training.") parser.add_argument( "-p", "--plugin", type=str, default="zero2", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d", "ddp", "moe"], help="Choose which plugin to use", ) parser.add_argument("--save_dir", type=str, default="checkpoint_dir", help="Checkpoint directory") parser.add_argument("--tensorboard_dir", type=str, default=None, help="Tensorboard directory") parser.add_argument("--config_file", type=str, default="training_config.json", help="Config file") # Training parameters parser.add_argument("-n", "--num_epochs", type=int, default=1, help="Number of training epochs") parser.add_argument("--accumulation_steps", type=int, default=1, help="Number of accumulation steps") parser.add_argument("--batch_size", type=int, default=2, help="Global Batch size of each process") parser.add_argument("--lr", type=float, default=3e-4, help="Learning rate") parser.add_argument("--max_length", type=int, default=8192, help="Model max length") parser.add_argument( "--mixed_precision", type=str, default="bf16", choices=["fp16", "bf16"], help="Mixed precision", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument( "-g", "--use_grad_checkpoint", action="store_true", default=False, help="Use gradient checkpointing", ) parser.add_argument( "-f", "--use_flash_attn", action="store_true", default=False, help="Use flash-attention", ) # Additional arguments for 3d plugin. parser.add_argument("--tp", type=int, default=1, help="TP size, used for 3d plugin.") parser.add_argument("--pp", type=int, default=1, help="PP size, used for 3d plugin.") parser.add_argument("--sp", type=int, default=1, help="SP size, used for 3d plugin.") parser.add_argument("--ep", type=int, default=1, help="EP size, used for moe plugin.") parser.add_argument("--zero_stage", type=int, default=1, help="Zero stage, used for 3d plugin.", choices=[0, 1, 2]) parser.add_argument( "--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"], help="SP mode, used for 3d plugin.", ) parser.add_argument( "--enable_sequence_parallelism", default=False, action="store_true", help="Whether to enable SP, used for 3d plugin.", ) parser.add_argument( "--zero_cpu_offload", default=False, action="store_true", help="Whether to use offloading, used for 3d plugin." ) parser.add_argument( "--microbatch_size", type=int, default=1, help="Batch size for each process in PP, used for 3d plugin." ) parser.add_argument("--lora_rank", type=int, default=0, help="lora rank when using lora to train.") parser.add_argument("--lora_alpha", type=int, default=8, help="lora alpha when using lora to train.") args = parser.parse_args() if args.plugin in ["3d", "moe"] and args.pp > 1 and args.accumulation_steps > 1: raise ValueError("Accumulation steps should be 1 when using PP. Please adjust batch size directly.") train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/training_scripts/train_rm.py
applications/ColossalChat/examples/training_scripts/train_rm.py
import argparse import json import math import os import resource from contextlib import nullcontext import torch from coati.dataset import DataCollatorForPreferenceDataset, StatefulDistributedSampler, load_tokenized_dataset from coati.models import LogExpLoss, LogSigLoss, LoraConfig, RewardModel, convert_to_lora_module from coati.trainer import RewardModelTrainer from coati.utils import load_checkpoint from transformers import AutoTokenizer import colossalai from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, HybridParallelPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.cluster import DistCoordinator from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import HybridAdam from colossalai.shardformer.policies.auto_policy import get_autopolicy logger = get_dist_logger() def train(args): lora_config = None if args.lora_config is not None: lora_config = LoraConfig.from_file(args.lora_config) # check lora compatibility if "gemini" in args.plugin and lora_config is not None and lora_config.r > 0: raise ValueError("LoRA is not supported in GeminiPlugin. Please use other plugin") if args.plugin == "gemini_auto" and args.accumulation_steps > 1: raise ValueError("Gradient accumulation is not supported in GeminiPlugin. Please use other plugin") # ============================== # Initialize Distributed Training # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # ====================================================== # Initialize Model, Objective, Optimizer and LR Scheduler # ====================================================== # Temp Fix: Disable lazy init due to version conflict # init_ctx = ( # LazyInitContext(default_device=get_current_device()) if isinstance(plugin, (GeminiPlugin,)) else nullcontext() # ) init_ctx = nullcontext() with init_ctx: if args.use_flash_attn: model = RewardModel( args.pretrain, torch_dtype=torch.bfloat16 if args.mixed_precision == "bf16" else torch.float16, use_flash_attention_2=True, ) coordinator.print_on_master(msg="Flash-attention enabled successfully") else: model = RewardModel( args.pretrain, ) if lora_config is not None: model = convert_to_lora_module(model, lora_config=lora_config) for name, module in model.named_modules(): if "norm" in name or "gate" in name: module = module.to(torch.float32) # ============================== # Initialize Booster # ============================== if args.plugin == "ddp": """ Default torch ddp plugin without any acceleration, for debugging purpose acceleration, for debugging purpose """ plugin = TorchDDPPlugin(find_unused_parameters=True) elif args.plugin == "gemini": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="static", initial_scale=2**16, max_norm=args.grad_clip, enable_flash_attention=args.use_flash_attn, enable_gradient_accumulation=True, ) elif args.plugin == "gemini_auto": plugin = GeminiPlugin( precision=args.mixed_precision, placement_policy="auto", initial_scale=2**16, enable_flash_attention=args.use_flash_attn, max_norm=args.grad_clip, ) elif args.plugin == "zero2": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, max_norm=args.grad_clip, ) elif args.plugin == "zero2_cpu": plugin = LowLevelZeroPlugin( stage=2, precision=args.mixed_precision, initial_scale=2**16, cpu_offload=True, max_norm=args.grad_clip, ) elif args.plugin == "3d": plugin = HybridParallelPlugin( tp_size=args.tp, pp_size=args.pp, sp_size=args.sp, sequence_parallelism_mode=args.sp_mode, zero_stage=args.zero_stage, enable_flash_attention=args.use_flash_attn, enable_sequence_parallelism=args.enable_sequence_parallelism, cpu_offload=True if args.zero_stage >= 1 and args.zero_cpu_offload else False, parallel_output=False, max_norm=args.grad_clip, precision=args.mixed_precision, custom_policy=get_autopolicy(model.model), ) else: raise ValueError(f"Unknown plugin {args.plugin}") booster = Booster(plugin=plugin) if args.grad_checkpoint: model.model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False}) coordinator.print_on_master(msg="Gradient checkpointing enabled successfully") # configure tokenizer tokenizer_dir = args.tokenizer_dir if args.tokenizer_dir is not None else args.pretrain tokenizer = AutoTokenizer.from_pretrained(tokenizer_dir, use_fast=False, trust_remote_code=True) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) tokenizer.padding_side = "right" tokenizer.add_bos_token = False tokenizer.add_eos_token = False # configure loss function if args.loss_fn == "log_sig": loss_fn = LogSigLoss() elif args.loss_fn == "log_exp": loss_fn = LogExpLoss() else: raise ValueError(f'Unsupported loss function "{args.loss_fn}"') # configure optimizer optim = HybridAdam( model_params=model.parameters(), lr=args.lr, betas=(0.9, 0.95), weight_decay=args.weight_decay, adamw_mode=True, ) # configure dataset coordinator.print_on_master(f"Load dataset: {args.dataset}") mode_map = {"train": "train", "valid": "validation", "test": "test"} train_dataset = load_tokenized_dataset(dataset_paths=args.dataset, mode="train", mode_map=mode_map) data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) train_dataloader = plugin.prepare_dataloader( dataset=train_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) eval_dataloader = None if args.eval_dataset: eval_dataset = load_tokenized_dataset(dataset_paths=args.eval_dataset, mode="dev") eval_data_collator = DataCollatorForPreferenceDataset(tokenizer=tokenizer, max_length=args.max_length) eval_dataloader = plugin.prepare_dataloader( dataset=eval_dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=eval_data_collator, distributed_sampler_cls=StatefulDistributedSampler, ) else: logger.warning("No evaluation dataset is provided, skip evaluation") num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps math.ceil(args.max_epochs * num_update_steps_per_epoch) if args.warmup_steps is None: args.warmup_steps = int(args.max_epochs * 0.025 * (len(train_dataloader) // args.accumulation_steps)) coordinator.print_on_master(f"Warmup steps is set to {args.warmup_steps}") lr_scheduler = CosineAnnealingWarmupLR( optimizer=optim, total_steps=args.max_epochs * num_update_steps_per_epoch, warmup_steps=args.warmup_steps, eta_min=0.1 * args.lr, ) default_dtype = torch.float16 if args.mixed_precision == "fp16" else torch.bfloat16 torch.set_default_dtype(default_dtype) model, optim, _, train_dataloader, lr_scheduler = booster.boost( model=model, optimizer=optim, lr_scheduler=lr_scheduler, dataloader=train_dataloader, ) torch.set_default_dtype(torch.float) coordinator.print_on_master(f"Booster init max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB") coordinator.print_on_master( f"Booster init max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) start_epoch = 0 sampler_start_idx = 0 start_step = 0 if args.checkpoint_path is not None: if "modeling" in args.checkpoint_path: coordinator.print_on_master(f"Continued pretrain from checkpoint {args.checkpoint_path}") booster.load_model(model, args.checkpoint_path) else: coordinator.print_on_master(f"Load model checkpoint from {args.checkpoint_path}") start_epoch, start_step, sampler_start_idx = load_checkpoint( load_dir=args.checkpoint_path, booster=booster, model=model, optimizer=optim, lr_scheduler=lr_scheduler, ) assert isinstance(train_dataloader.sampler, StatefulDistributedSampler) train_dataloader.sampler.set_start_index(start_index=sampler_start_idx) coordinator.print_on_master( f"Loaded checkpoint {args.checkpoint_path} at epoch {start_epoch} step {start_step}" ) coordinator.print_on_master(f"Loaded sample at index {sampler_start_idx}") coordinator.print_on_master( f"Checkpoint loaded max CUDA memory: {torch.cuda.max_memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded CUDA memory: {torch.cuda.memory_allocated() / 1024 ** 2:.2f} MB" ) coordinator.print_on_master( f"Checkpoint loaded max CPU memory: {resource.getrusage(resource.RUSAGE_SELF).ru_maxrss / 1024:.2f} MB" ) trainer = RewardModelTrainer( model, booster, optim, plugin, lr_scheduler, tokenizer, loss_fn=loss_fn, max_epochs=args.max_epochs, accumulation_steps=args.accumulation_steps, start_epoch=start_epoch, save_interval=args.save_interval, save_dir=args.save_dir, coordinator=coordinator, ) trainer.fit( train_preference_dataloader=train_dataloader, eval_preference_dataloader=eval_dataloader, log_dir=args.log_dir, use_wandb=args.use_wandb, ) if lora_config is not None and lora_config.r > 0: # NOTE: set model to eval to merge LoRA weights model.eval() # save model checkpoint after fitting on only rank0 if args.save_dir is not None: coordinator.print_on_master("Start saving final model checkpoint") booster.save_model(model, os.path.join(args.save_dir, "modeling"), shard=True) coordinator.print_on_master( f"Saved final model checkpoint at epoch {args.max_epochs} at folder {args.save_dir}" ) coordinator.print_on_master(f"Max CUDA memory usage: {torch.cuda.max_memory_allocated()/1024**2:.2f} MB") if __name__ == "__main__": # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument( "--plugin", type=str, default="gemini", choices=["gemini", "gemini_auto", "zero2", "zero2_cpu", "3d", "ddp"], help="Choose which plugin to use", ) parser.add_argument("--grad_clip", type=float, default=1.0, help="Gradient clipping value") parser.add_argument("--weight_decay", type=float, default=0.1, help="Weight decay") parser.add_argument("--warmup_steps", type=int, default=None, help="Warmup steps") parser.add_argument("--tp", type=int, default=1) parser.add_argument("--pp", type=int, default=1) parser.add_argument("--sp", type=int, default=1) parser.add_argument("--enable_sequence_parallelism", default=False, action="store_true") parser.add_argument("--zero_stage", type=int, default=0, help="Zero stage", choices=[0, 1, 2]) parser.add_argument("--zero_cpu_offload", default=False, action="store_true") parser.add_argument("--sp_mode", type=str, default="split_gather", choices=["split_gather", "ring", "all_to_all"]) parser.add_argument("--pretrain", type=str, default=None) parser.add_argument("--tokenizer_dir", type=str, default=None) parser.add_argument("--dataset", nargs="+", default=[]) parser.add_argument("--eval_dataset", nargs="+", default=[]) parser.add_argument( "--checkpoint_path", type=str, default=None, help="Checkpoint path if need to resume training form a checkpoint" ) parser.add_argument("--config_file", type=str, default=None, help="Config file") parser.add_argument("--save_dir", type=str, default=None) parser.add_argument("--max_length", type=int, default=2048, help="Model max length") parser.add_argument("--max_epochs", type=int, default=3) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--mixed_precision", type=str, default="fp16", choices=["fp16", "bf16"], help="Mixed precision") parser.add_argument("--loss_fn", type=str, default="log_sig", choices=["log_sig", "log_exp"], help="Loss function") parser.add_argument("--lora_config", type=str, default=None, help="low-rank adaptation config file path") parser.add_argument("--save_interval", type=int, default=1000, help="number of step between two checkpoints") parser.add_argument("--lr", type=float, default=5e-6) parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--log_dir", default=None, type=str) parser.add_argument("--use_wandb", default=False, action="store_true") parser.add_argument("--grad_checkpoint", default=False, action="store_true") parser.add_argument("--use_flash_attn", default=False, action="store_true") args = parser.parse_args() if args.config_file is not None: os.makedirs(os.path.dirname(args.config_file), exist_ok=True) with open(args.config_file, "w") as f: json.dump(args.__dict__, f, indent=4) train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/inference/inference.py
applications/ColossalChat/examples/inference/inference.py
import argparse import json import os from typing import Dict import torch from chatio import dummy_io, rich_io, simple_io from coati.dataset.conversation import setup_conversation_template from coati.models import generate_streaming from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedModel from colossalai.logging import get_dist_logger logger = get_dist_logger() def get_gpu_memory(max_gpus=None): """ Get the available memory for each GPU. Args: max_gpus (int, optional): The maximum number of GPUs to consider. Defaults to None. Returns: list: A list of available memory for each GPU. """ gpu_memory = [] num_gpus = torch.cuda.device_count() if max_gpus is None else min(max_gpus, torch.cuda.device_count()) for gpu_id in range(num_gpus): # Code to get GPU memory goes here with torch.cuda.device(gpu_id): device = torch.cuda.current_device() gpu_properties = torch.cuda.get_device_properties(device) total_memory = gpu_properties.total_memory / (1024**3) allocated_memory = torch.cuda.memory_allocated() / (1024**3) available_memory = total_memory - allocated_memory gpu_memory.append(available_memory) return gpu_memory def load_model_and_tokenizer(model_path, tokenizer_path, device="cuda", **kwargs): """ Load the model and tokenizer from the specified paths and move the model to the specified device. Args: model_path (str): The path to the pre-trained model. tokenizer_path (str): The path to the pre-trained tokenizer. device (str, optional): The device to move the model to. Defaults to "cuda". **kwargs: Additional keyword arguments to be passed to the `AutoModelForCausalLM.from_pretrained` function. Returns: tuple: A tuple containing the loaded model and tokenizer. """ model = AutoModelForCausalLM.from_pretrained(model_path, **kwargs, trust_remote_code=True).to(torch.bfloat16) tokenizer = AutoTokenizer.from_pretrained(tokenizer_path, trust_remote_code=True) tokenizer.pad_token = tokenizer.eos_token model.to(device) return model, tokenizer def _set_default_generate_kwargs(model: PreTrainedModel) -> Dict: """ Set default keyword arguments for generation based on the given model. Args: model (PreTrainedModel): The model used for generation. Returns: Dict: A dictionary containing the default keyword arguments for generation. """ unwrapped_model = model new_kwargs = {} # Use huggingface models method directly if hasattr(unwrapped_model, "prepare_inputs_for_generation"): new_kwargs["prepare_inputs_fn"] = unwrapped_model.prepare_inputs_for_generation if hasattr(unwrapped_model, "_update_model_kwargs_for_generation"): new_kwargs["update_model_kwargs_fn"] = unwrapped_model._update_model_kwargs_for_generation return new_kwargs def generation_wrapper(*args, **kwargs): input_ids = args[1] tokenizer = args[2] for output in generate_streaming(*args, **kwargs): yield tokenizer.batch_decode(output[:, input_ids.size(1) :], skip_special_tokens=True)[0] def main(args): conversation_template_config = json.load(open(args.conversation_template_config, "r", encoding="utf8")) max_new_tokens = args.max_new_tokens model_max_length = args.model_max_length model, tokenizer = load_model_and_tokenizer( args.model_path, args.tokenizer_path or args.model_path, local_files_only=True ) assert max_new_tokens <= model_max_length if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") tokenizer.padding_side = "left" model_kwargs = { "max_new_tokens": max_new_tokens, # 'early_stopping': True, # 'top_k': -1, # 'top_p': 1.0, # 'temperature': 1.0, # 'temperature':0.1, } round = 1 conv = setup_conversation_template(tokenizer, conversation_template_config, args.conversation_template_config) while True: if args.io == "simple": chat_io = simple_io elif args.io == "rich": chat_io = rich_io elif args.io == "dummy": chat_io = dummy_io else: raise ValueError(f"Unknown io type: {args.io}") # raw_text = print(">>> Human:", end=" ") inp = chat_io.prompt_for_input("user") if not inp: print("prompt should not be empty!") continue if inp.strip() == "clear": conv.clear() os.system("clear") continue if inp.strip() == "exit": print("End of chat.") break query_text = inp.strip() conv.append_message("user", query_text) chat_io.prompt_for_output("assistant") prompt = conv.get_prompt(add_generation_prompt=True) input_ids = tokenizer(prompt, return_tensors="pt", add_special_tokens=False)["input_ids"].to( torch.cuda.current_device() ) default_generate_kwargs = _set_default_generate_kwargs(model) model_kwargs.update(default_generate_kwargs) output_stream = generation_wrapper( model, input_ids, tokenizer, max_length=model_max_length, temperature=0.7, early_stopping=True, stop_token_ids=conversation_template_config["stop_ids"], **model_kwargs, ) # print(f">>> Assistant:", end=" ") outputs = chat_io.stream_output(output_stream) conv.append_message("assistant", outputs.strip()) with open("round.txt", mode="a", encoding="utf-8") as f: f.write("\n\n" + "=" * 10 + "\n") f.write(f"round {round}:\n{conv.save_prompt()}\n\n") f.write("=" * 10 + "\n") # print(f">>> Assistant:", end=" ") round += 1 if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--model_path", type=str, default=None) parser.add_argument("--tokenizer_path", type=str, default=None) parser.add_argument("--conversation_template_config", type=str, default=None) parser.add_argument("--model_max_length", type=int, default=2048) parser.add_argument("--max_new_tokens", type=int, default=512) parser.add_argument("--io", type=str, default="rich", choices=["simple", "rich", "dummy"]) args = parser.parse_args() main(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/inference/chatio.py
applications/ColossalChat/examples/inference/chatio.py
""" command line IO utils for chatbot """ import abc import re from prompt_toolkit import PromptSession from prompt_toolkit.auto_suggest import AutoSuggestFromHistory from prompt_toolkit.completion import WordCompleter from prompt_toolkit.history import InMemoryHistory from rich.console import Console from rich.live import Live from rich.markdown import Markdown class ChatIO(abc.ABC): @abc.abstractmethod def prompt_for_input(self, role: str) -> str: """Prompt for input from a role.""" @abc.abstractmethod def prompt_for_output(self, role: str): """Prompt for output from a role.""" @abc.abstractmethod def stream_output(self, output_stream): """Stream output.""" class SimpleChatIO(ChatIO): def prompt_for_input(self, role) -> str: return input(f"{role}: ") def prompt_for_output(self, role: str): print(f"{role}: ", end="", flush=True) def stream_output(self, output_stream): pre = 0 for outputs in output_stream: outputs = outputs.strip() outputs = outputs.split(" ") now = len(outputs) - 1 if now > pre: print(" ".join(outputs[pre:now]), end=" ", flush=True) pre = now print(" ".join(outputs[pre:]), flush=True) return " ".join(outputs) class RichChatIO(ChatIO): def __init__(self): self._prompt_session = PromptSession(history=InMemoryHistory()) self._completer = WordCompleter(words=["!exit", "!reset"], pattern=re.compile("$")) self._console = Console() def prompt_for_input(self, role) -> str: self._console.print(f"[bold]{role}:") prompt_input = self._prompt_session.prompt( completer=self._completer, multiline=False, auto_suggest=AutoSuggestFromHistory(), key_bindings=None, ) self._console.print() return prompt_input def prompt_for_output(self, role: str) -> str: self._console.print(f"[bold]{role}:") def stream_output(self, output_stream): """Stream output from a role.""" # Create a Live context for updating the console output with Live(console=self._console, refresh_per_second=60) as live: # Read lines from the stream for outputs in output_stream: accumulated_text = outputs if not accumulated_text: continue # Render the accumulated text as Markdown # NOTE: this is a workaround for the rendering "unstandard markdown" # in rich. The chatbots output treat "\n" as a new line for # better compatibility with real-world text. However, rendering # in markdown would break the format. It is because standard markdown # treat a single "\n" in normal text as a space. # Our workaround is adding two spaces at the end of each line. # This is not a perfect solution, as it would # introduce trailing spaces (only) in code block, but it works well # especially for console output, because in general the console does not # care about trailing spaces. lines = [] for line in accumulated_text.splitlines(): lines.append(line) if line.startswith("```"): # Code block marker - do not add trailing spaces, as it would # break the syntax highlighting lines.append("\n") else: lines.append(" \n") markdown = Markdown("".join(lines)) # Update the Live console output live.update(markdown) self._console.print() return outputs class DummyChatIO(ChatIO): """ Dummy ChatIO class for testing """ def __init__(self): self.roles = [] self._console = Console() def prompt_for_input(self, role) -> str: self.roles.append(role) if len(self.roles) == 1: ret = "Hello" elif len(self.roles) == 2: ret = "What's the value of 1+1?" else: ret = "exit" self._console.print(f"[bold]{role}:{ret}") return ret def prompt_for_output(self, role: str) -> str: self._console.print(f"[bold]{role}:") def stream_output(self, output_stream): """Stream output from a role.""" # Create a Live context for updating the console output with Live(console=self._console, refresh_per_second=60) as live: # Read lines from the stream for outputs in output_stream: accumulated_text = outputs if not accumulated_text: continue # Render the accumulated text as Markdown # NOTE: this is a workaround for the rendering "unstandard markdown" # in rich. The chatbots output treat "\n" as a new line for # better compatibility with real-world text. However, rendering # in markdown would break the format. It is because standard markdown # treat a single "\n" in normal text as a space. # Our workaround is adding two spaces at the end of each line. # This is not a perfect solution, as it would # introduce trailing spaces (only) in code block, but it works well # especially for console output, because in general the console does not # care about trailing spaces. lines = [] for line in accumulated_text.splitlines(): lines.append(line) if line.startswith("```"): # Code block marker - do not add trailing spaces, as it would # break the syntax highlighting lines.append("\n") else: lines.append(" \n") markdown = Markdown("".join(lines)) # Update the Live console output live.update(markdown) self._console.print() return outputs simple_io = SimpleChatIO() rich_io = RichChatIO() dummy_io = DummyChatIO()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/inference/web_chatbot/utils.py
applications/ColossalChat/examples/inference/web_chatbot/utils.py
import copy import json from threading import Lock from typing import List import jieba import torch from coati.dataset.conversation import default_conversation from pydantic import BaseModel, Field def update_model_kwargs_fn(outputs: dict, **model_kwargs) -> dict: if "past_key_values" in outputs: model_kwargs["past"] = outputs["past_key_values"] else: model_kwargs["past"] = None # update token_type_ids with last value if "token_type_ids" in model_kwargs: token_type_ids = model_kwargs["token_type_ids"] model_kwargs["token_type_ids"] = torch.cat([token_type_ids, token_type_ids[:, -1].unsqueeze(-1)], dim=-1) # update attention mask if "attention_mask" in model_kwargs: attention_mask = model_kwargs["attention_mask"] model_kwargs["attention_mask"] = torch.cat( [attention_mask, attention_mask.new_ones((attention_mask.shape[0], 1))], dim=-1 ) return model_kwargs class Dialogue(BaseModel): instruction: str = Field(min_length=1, example="Count up from 1 to 500.") response: str = Field(example="") class ChatPromptProcessor: SAFE_RESPONSE = "The input/response contains inappropriate content, please rephrase your prompt." def __init__(self, censored_words: List[str] = []): self.censored_words = set([word.lower() for word in censored_words]) self.conv = copy.deepcopy(default_conversation) def preprocess_prompt(self, history: List[Dialogue]) -> str: self.conv.clear() for round in history: self.conv.append_message(self.conv.roles[0], round.instruction) if len(round.instruction) > 0: self.conv.append_message(self.conv.roles[1], round.response) return self.conv.get_prompt() def postprocess_output(self, output: str) -> str: return output.strip() def has_censored_words(self, text: str) -> bool: if len(self.censored_words) == 0: return False intersection = set(jieba.cut(text.lower())) & self.censored_words return len(intersection) > 0 class LockedIterator: def __init__(self, it, lock: Lock) -> None: self.lock = lock self.it = iter(it) def __iter__(self): return self def __next__(self): with self.lock: return next(self.it) def load_json(path: str): with open(path) as f: return json.load(f)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/inference/web_chatbot/locustfile.py
applications/ColossalChat/examples/inference/web_chatbot/locustfile.py
from locust import HttpUser, task samples = [ [ dict( instruction="Who is the best player in the history of NBA?", response="The best player in the history of the NBA is widely considered to be Michael Jordan. He is one of the most successful players in the league, having won 6 NBA championships with the Chicago Bulls and 5 more with the Washington Wizards. He is a 5-time MVP, 1", ), dict(instruction="continue this talk", response=""), ], [ dict(instruction="Who is the best player in the history of NBA?", response=""), ], ] class GenerationUser(HttpUser): @task def generate(self): for sample in samples: data = {"max_new_tokens": 64, "history": sample} with self.client.post("/generate", json=data, catch_response=True) as response: if response.status_code in (200, 406): response.success() else: response.failure("Response wrong")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/inference/web_chatbot/server.py
applications/ColossalChat/examples/inference/web_chatbot/server.py
import argparse import os from threading import Lock from typing import Generator, List, Optional import torch import uvicorn from coati.models import generate_streaming from coati.quant import llama_load_quant, low_resource_init from fastapi import FastAPI, Request from fastapi.middleware.cors import CORSMiddleware from pydantic import BaseModel, Field from slowapi import Limiter, _rate_limit_exceeded_handler from slowapi.errors import RateLimitExceeded from slowapi.util import get_remote_address from sse_starlette.sse import EventSourceResponse from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer from utils import ChatPromptProcessor, Dialogue, LockedIterator, load_json, update_model_kwargs_fn MAX_LEN = 512 running_lock = Lock() class GenerationTaskReq(BaseModel): max_new_tokens: int = Field(gt=0, le=512, example=64) history: List[Dialogue] = Field(min_items=1) top_k: Optional[int] = Field(default=None, gt=0, example=50) top_p: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.5) temperature: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.7) repetition_penalty: Optional[float] = Field(default=None, gt=1.0, example=1.2) limiter = Limiter(key_func=get_remote_address) app = FastAPI() app.state.limiter = limiter app.add_exception_handler(RateLimitExceeded, _rate_limit_exceeded_handler) # set CORS origin_spec_from_env = os.environ.get("CORS_ORIGIN", None) if origin_spec_from_env is not None: # allow CORS from the specified origins origins = os.environ["CORS_ORIGIN"].split(",") else: # allow CORS from all origins origins = ["*"] app.add_middleware( CORSMiddleware, allow_origins=origins, allow_credentials=True, allow_methods=["*"], allow_headers=["*"], ) def generate_streamingly(prompt, max_length, max_new_tokens, top_k, top_p, temperature): input_ids = tokenizer(prompt, return_tensors="pt")["input_ids"] # TODO(ver217): streaming generation does not support repetition_penalty now model_kwargs = { "max_new_tokens": max_new_tokens, "early_stopping": True, "top_k": top_k, "top_p": top_p, "temperature": temperature, "prepare_inputs_fn": None, "update_model_kwargs_fn": update_model_kwargs_fn, } is_first_word = True generator = LockedIterator( generate_streaming(model, input_ids, tokenizer, max_length, **model_kwargs), running_lock ) for output in generator: output = output.cpu() tokens = tokenizer.convert_ids_to_tokens(output, skip_special_tokens=True) current_sub_tokens = [] for token in tokens: if token in tokenizer.all_special_tokens: continue current_sub_tokens.append(token) if current_sub_tokens: out_string = tokenizer.sp_model.decode(current_sub_tokens) if is_first_word: out_string = out_string.lstrip() is_first_word = False elif current_sub_tokens[0].startswith("▁"): # whitespace will be ignored by the frontend out_string = " " + out_string yield out_string async def event_generator(request: Request, generator: Generator): while True: if await request.is_disconnected(): break try: yield {"event": "generate", "data": next(generator)} except StopIteration: yield {"event": "end", "data": ""} break @app.post("/generate/stream") @limiter.limit("1/second") def generate(data: GenerationTaskReq, request: Request): prompt = prompt_processor.preprocess_prompt(data.history) event_source = event_generator( request, generate_streamingly(prompt, data.max_length, data.max_new_tokens, data.top_k, data.top_p, data.temperature), ) return EventSourceResponse(event_source) @app.post("/generate") @limiter.limit("1/second") def generate_no_stream(data: GenerationTaskReq, request: Request): prompt = prompt_processor.preprocess_prompt(data.history, data.max_new_tokens) if prompt_processor.has_censored_words(prompt): return prompt_processor.SAFE_RESPONSE inputs = {k: v.cuda() for k, v in tokenizer(prompt, return_tensors="pt").items()} with running_lock: output = model.generate(**inputs, **data.dict(exclude={"history"})) output = output.cpu() prompt_len = inputs["input_ids"].size(1) response = output[0, prompt_len:] out_string = tokenizer.decode(response, skip_special_tokens=True) out_string = prompt_processor.postprocess_output(out_string) if prompt_processor.has_censored_words(out_string): return prompt_processor.SAFE_RESPONSE return out_string if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "pretrained", help="Path to pretrained model. Can be a local path or a model name from the HuggingFace model hub.", ) parser.add_argument( "--tokenizer_path", help="Path to pretrained tokenizer. Can be a local path or a model name from the HuggingFace model hub.", default=None, ) parser.add_argument( "--quant", choices=["8bit", "4bit"], default=None, help="Quantization mode. Default: None (no quantization, fp16).", ) parser.add_argument( "--gptq_checkpoint", default=None, help="Path to GPTQ checkpoint. This is only useful when quantization mode is 4bit. Default: None.", ) parser.add_argument( "--gptq_group_size", type=int, default=128, help="Group size for GPTQ. This is only useful when quantization mode is 4bit. Default: 128.", ) parser.add_argument("--http_host", default="0.0.0.0") parser.add_argument("--http_port", type=int, default=7070) parser.add_argument( "--profanity_file", default=None, help="Path to profanity words list. It should be a JSON file containing a list of words.", ) args = parser.parse_args() if args.quant == "4bit": assert args.gptq_checkpoint is not None, "Please specify a GPTQ checkpoint." if args.tokenizer_path is None: args.tokenizer_path = args.pretrained tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_path, local_files_only=True) if args.profanity_file is not None: censored_words = load_json(args.profanity_file) else: censored_words = [] prompt_processor = ChatPromptProcessor(censored_words=censored_words) if args.quant == "4bit": with low_resource_init(): config = AutoConfig.from_pretrained(args.pretrained) model = AutoModelForCausalLM(config) model = llama_load_quant(model, args.gptq_checkpoint, 4, args.gptq_group_size) model.cuda() else: model = AutoModelForCausalLM.from_pretrained( args.pretrained, load_in_8bit=(args.quant == "8bit"), torch_dtype=torch.float16, device_map="auto", local_files_only=True, ) if args.quant != "8bit": model.half() # seems to fix bugs for some users. model.eval() config = uvicorn.Config(app, host=args.http_host, port=args.http_port) server = uvicorn.Server(config=config) server.run() """ python server.py /home/lcyab/data/models/experiments5/checkpoint/experiment5-2023-10-20-21-53-51/modeling/ --tokenizer_path /mnt/vepfs/lcxyc/leaderboard_models/Colossal-LLaMA-2-7b-base/ """
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/applications/ColossalChat/examples/data_preparation_scripts/prepare_dataset.py
applications/ColossalChat/examples/data_preparation_scripts/prepare_dataset.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Prepare dataset scripts Usage: - For SFT dataset preparation (SFT) python prepare_dataset.py --type sft \ --data_input_dirs /PATH/TO/SFT/DATASET \ --conversation_template_config /PATH/TO/CHAT/TEMPLATE/CONFIG.json \ --tokenizer_dir "" \ --data_cache_dir $SAVE_DIR/cache \ --data_jsonl_output_dir $SAVE_DIR/jsonl \ --data_arrow_output_dir $SAVE_DIR/arrow \ - For prompt dataset preparation (PPO) python prepare_dataset.py --type prompt \ --data_input_dirs /PATH/TO/SFT/DATASET \ --conversation_template_config /PATH/TO/CHAT/TEMPLATE/CONFIG.json \ --tokenizer_dir "" \ --data_cache_dir $SAVE_DIR/cache \ --data_jsonl_output_dir $SAVE_DIR/jsonl \ --data_arrow_output_dir $SAVE_DIR/arrow \ - For Preference dataset preparation (DPO and Reward model training) python prepare_dataset.py --type preference \ --data_input_dirs /PATH/TO/SFT/DATASET \ --conversation_template_config /PATH/TO/CHAT/TEMPLATE/CONFIG.json \ --tokenizer_dir "" \ --data_cache_dir $SAVE_DIR/cache \ --data_jsonl_output_dir $SAVE_DIR/jsonl \ --data_arrow_output_dir $SAVE_DIR/arrow \ """ import argparse import json import math import os import random import time from multiprocessing import cpu_count from coati.dataset import setup_conversation_template, tokenize_kto, tokenize_prompt, tokenize_rlhf, tokenize_sft from datasets import dataset_dict, load_dataset from transformers import AutoTokenizer from colossalai.logging import get_dist_logger logger = get_dist_logger() def main(): parser = argparse.ArgumentParser() parser.add_argument( "--type", type=str, required=True, default=None, choices=["sft", "prompt", "preference", "kto"], help="Type of dataset, chose from 'sft', 'prompt', 'preference'. 'kto'", ) parser.add_argument( "--data_input_dirs", type=str, required=True, default=None, help="Comma(i.e., ',') separated list of all data directories containing `.jsonl` data files.", ) parser.add_argument( "--tokenizer_dir", type=str, required=True, default=None, help="A directory containing the tokenizer" ) parser.add_argument( "--conversation_template_config", type=str, default="conversation_template_config", help="Path to save conversation template config files.", ) parser.add_argument("--data_cache_dir", type=str, default="cache", help="Data cache directory") parser.add_argument( "--data_jsonl_output_dir", type=str, default="jsonl_output", help="Output directory of spliced dataset with jsonl format", ) parser.add_argument( "--data_arrow_output_dir", type=str, default="arrow_output", help="Output directory of spliced dataset with arrow format", ) parser.add_argument("--max_length", type=int, default=4096, help="Max length of each spliced tokenized sequence") parser.add_argument("--num_spliced_dataset_bins", type=int, default=10, help="Number of spliced dataset bins") parser.add_argument( "--num_samples_per_datafile", type=int, default=-1, help="Number of samples to be generated from each data file. -1 denote all samples.", ) args = parser.parse_args() if args.num_spliced_dataset_bins >= 100000: raise ValueError("Too many spliced divisions, must be smaller than 100000") assert not os.path.exists(args.data_cache_dir), f"Find existed data cache dir {args.data_cache_dir}" assert not os.path.exists( args.data_jsonl_output_dir ), f"Find existed jsonl data output dir {args.data_jsonl_output_dir}" assert not os.path.exists( args.data_arrow_output_dir ), f"Find existed arrow data output dir {args.data_arrow_output_dir}" os.makedirs(args.data_jsonl_output_dir) os.makedirs(args.data_arrow_output_dir) # Prepare to all input datasets input_data_paths = [] input_data_dirs = args.data_input_dirs.split(",") for ds_dir in input_data_dirs: ds_dir = os.path.abspath(ds_dir) assert os.path.exists(ds_dir), f"Not find data dir {ds_dir}" ds_files = [name for name in os.listdir(ds_dir) if name.endswith(".jsonl")] ds_paths = [os.path.join(ds_dir, name) for name in ds_files] input_data_paths.extend(ds_paths) # Prepare to data splitting. train_splits = [] split_interval = math.ceil(100 / args.num_spliced_dataset_bins) for i in range(0, 100, split_interval): start = i end = i + split_interval if end > 100: end = 100 train_splits.append(f"train[{start}%:{end}%]") # Prepare the tokenizer. tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_dir, use_fast=False, trust_remote_code=True) if os.path.exists(args.conversation_template_config): chat_template_config = json.load(open(args.conversation_template_config, "r", encoding="utf8")) else: chat_template_config = { "system_message": "A chat between a curious human and an artificial intelligence assistant. " "The assistant gives helpful, detailed, and polite answers to the human's questions.\n\n" } # Use default system message if args.type == "preference": if "stop_ids" not in chat_template_config: # Ask the user to define stop_ids for PPO training dummy_messages = [ {"role": "user", "content": "Hello, how are you?"}, {"role": "assistant", "content": "I'm doing great. How can I help you today?"}, {"role": "user", "content": "Who made you?"}, {"role": "assistant", "content": "I am a chatbot trained by Colossal-AI."}, ] dummy_prompt = tokenizer.apply_chat_template(dummy_messages, tokenize=False) tokenized = tokenizer(dummy_prompt, add_special_tokens=False)["input_ids"] tokens = tokenizer.convert_ids_to_tokens(tokenized, skip_special_tokens=False) corresponding_str = [tokenizer.convert_tokens_to_string([token]) for token in tokens] token_id_mapping = [{"token": s, "id": tokenized[i]} for i, s in enumerate(corresponding_str)] stop_ids = input( "For PPO, we recommend to provide stop_ids for the properly stop the generation during roll out stage. " "stop_ids are the ids of repetitive pattern that indicate the end of the assistant's response. " "Here is an example of formatted prompt and token-id mapping, you can set stop_ids by entering a list " "of integers, separate by space, press `Enter` to end. Or you can press `Enter` without input if you are " "not using PPO or you prefer to not set the stop_ids, in that case, stop_ids will be set to tokenizer.eos_token_id. " f"\nPrompt:\n{dummy_prompt}\nToken-id Mapping:\n{token_id_mapping}\nstop_ids:" ) if stop_ids == "": chat_template_config["stop_ids"] = [tokenizer.eos_token_id] else: try: chat_template_config["stop_ids"] = [int(s) for s in stop_ids.split()] except ValueError: raise ValueError("Invalid input, please provide a list of integers.") else: # Set stop_ids to eos_token_id for other dataset types if not exist if "stop_ids" not in chat_template_config: chat_template_config["stop_ids"] = [tokenizer.eos_token_id] conversation_template = setup_conversation_template( tokenizer, chat_template_config=chat_template_config, save_path=args.conversation_template_config ) if hasattr(tokenizer, "pad_token") and hasattr(tokenizer, "eos_token") and tokenizer.eos_token is not None: try: # Some tokenizers doesn't allow to set pad_token mannually e.g., Qwen tokenizer.pad_token = tokenizer.eos_token except AttributeError as e: logger.warning(f"Unable to set pad token to eos token, {str(e)}") if not hasattr(tokenizer, "pad_token") or tokenizer.pad_token is None: logger.warning( "The tokenizer does not have a pad token which is required. May lead to unintended behavior in training, Please consider manually set them." ) list_dataset = load_dataset( path="json", data_files=input_data_paths, cache_dir=os.path.join(args.data_cache_dir, "raw"), keep_in_memory=False, split=train_splits, num_proc=cpu_count(), ) if args.type == "sft": preparation_function = tokenize_sft elif args.type == "prompt": preparation_function = tokenize_prompt elif args.type == "preference": preparation_function = tokenize_rlhf elif args.type == "kto": preparation_function = tokenize_kto else: raise ValueError("Unknow dataset type. Please choose one from ['sft', 'prompt', 'preference']") for index, dataset in enumerate(list_dataset): assert isinstance(dataset, dataset_dict.Dataset) if len(dataset) == 0: # Hack: Skip empty dataset. If dataset contains less than num_of_rank samples, some rank may have empty dataset and leads to error continue if args.num_samples_per_datafile > 0: # limit the number of samples in each dataset dataset = dataset.select( random.sample(range(len(dataset)), min(args.num_samples_per_datafile, len(dataset))) ) logger.info(f"Start to process part-{index}/{len(list_dataset)} of all original datasets.") dataset = dataset.map( function=preparation_function, fn_kwargs={ "tokenizer": tokenizer, "conversation_template": conversation_template, "max_length": args.max_length, }, keep_in_memory=False, num_proc=min(len(dataset), cpu_count()), ) if args.type == "kto": filter_by = "completion" elif args.type == "preference": filter_by = "chosen_input_ids" else: filter_by = "input_ids" dataset = dataset.filter(lambda data: data[filter_by] is not None) # Save each jsonl spliced dataset. output_index = "0" * (5 - len(str(index))) + str(index) output_name = f"part-{output_index}" output_jsonl_path = os.path.join(args.data_jsonl_output_dir, output_name + ".jsonl") st = time.time() with open(file=output_jsonl_path, mode="w", encoding="utf-8") as fp_writer: count = 0 for data_point in dataset: if count % 500 == 0: logger.info(f"processing {count} spliced data points for {fp_writer.name}") count += 1 fp_writer.write(json.dumps(data_point, ensure_ascii=False) + "\n") logger.info( f"Current file {fp_writer.name}; " f"Data size: {len(dataset)}; " f"Time cost: {round((time.time() - st) / 60, 6)} minutes." ) # Save each arrow spliced dataset output_arrow_path = os.path.join(args.data_arrow_output_dir, output_name) logger.info(f"Start to save {output_arrow_path}") dataset = load_dataset( path="json", data_files=[output_jsonl_path], cache_dir=os.path.join(args.data_cache_dir, "tokenized"), keep_in_memory=False, num_proc=cpu_count(), split="train", ) dataset.save_to_disk(dataset_path=output_arrow_path, num_proc=min(len(dataset), cpu_count())) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/__init__.py
colossalai/__init__.py
from . import accelerator from .initialize import launch, launch_from_openmpi, launch_from_slurm, launch_from_torch try: # .version will be created by setup.py from .version import __version__ except ModuleNotFoundError: # this will only happen if the user did not run `pip install` # and directly set PYTHONPATH to use Colossal-AI which is a bad practice __version__ = "0.0.0" print("please install Colossal-AI from https://www.colossalai.org/download or from source") __all__ = ["launch", "launch_from_openmpi", "launch_from_slurm", "launch_from_torch", "__version__"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/initialize.py
colossalai/initialize.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import os # set CUDA_DEVICE_MAX_CONNECTIONS=1 to ensure that when overlapping communication and computation, # the order of of kernel launches on GPUs are the same as on the CPU so that comm is launched first. # see https://github.com/NVIDIA/Megatron-LM/issues/533 # https://forums.developer.nvidia.com/t/how-many-streams-maximum-number-of-streams/6571/16 os.environ["CUDA_DEVICE_MAX_CONNECTIONS"] = "1" import torch import torch.distributed as dist from colossalai.accelerator import get_accelerator from colossalai.logging import get_dist_logger from colossalai.utils import set_seed def launch( rank: int, world_size: int, host: str, port: int, backend: str = "nccl", local_rank: int = None, seed: int = 1024, verbose: bool = True, ): """This function first parses the configuration arguments, using :func:`parse_args()` in case one of the input arguments are not given. Then initialize and set distributed environment by calling global_context's functions. Args: config (Union[str, dict, Config]): Config file or config file path are both acceptable rank (int): Rank for the default process group world_size (int): World size of the default process group host (str): The master address for distributed training port (str): The master port for distributed training backend (str, optional): Backend for ``torch.distributed``, defaults to ``nccl`` local_rank (int, optional): Rank for the process on the node and is used to set the default CUDA device, defaults to None. If local_rank = None, the default device ordinal will be calculated automatically. seed (int, optional): Specified random seed for every process. Defaults to 1024. verbose (bool, optional): Whether to print logs. Defaults to True. Raises: Exception: Raise exception when config type is wrong """ cur_accelerator = get_accelerator() backend = cur_accelerator.communication_backend # init default process group if ":" in host: # IPv6 init_method = f"tcp://[{host}]:{port}" else: # IPv4 init_method = f"tcp://{host}:{port}" dist.init_process_group(rank=rank, world_size=world_size, backend=backend, init_method=init_method) # set cuda device # if local rank is not given, calculate automatically if cur_accelerator.support_set_device: cur_accelerator.set_device(local_rank) set_seed(seed) try: torch._dynamo.config.optimize_ddp = world_size > 1 except AttributeError: pass if verbose: logger = get_dist_logger() logger.info(f"Distributed environment is initialized, world size: {dist.get_world_size()}", ranks=[0]) def launch_from_slurm( host: str, port: int, backend: str = "nccl", seed: int = 1024, verbose: bool = True, ): """A wrapper for colossalai.launch for SLURM launcher by reading rank and world size from the environment variables set by SLURM Args: config (Union[str, dict, Config]): Config file or config file path are both acceptable host (str): The master address for distributed training port (str): The master port for distributed training backend (str, optional): Backend for ``torch.distributed``, defaults to ``nccl`` seed (int, optional): Specified random seed for every process. Defaults to 1024. verbose (bool, optional): Whether to print logs. Defaults to True. """ try: rank = int(os.environ["SLURM_PROCID"]) world_size = int(os.environ["SLURM_NPROCS"]) except KeyError as e: raise RuntimeError( f"Could not find {e} in the SLURM environment, visit https://www.colossalai.org/ for more information on launching with SLURM" ) launch( rank=rank, world_size=world_size, host=host, port=port, backend=backend, seed=seed, verbose=verbose, ) def launch_from_openmpi( host: str, port: int, backend: str = "nccl", seed: int = 1024, verbose: bool = True, ): """A wrapper for colossalai.launch for OpenMPI launcher by reading rank and world size from the environment variables set by OpenMPI Args: config (Union[str, dict, Config]): Config file or config file path are both acceptable host (str): The master address for distributed training port (str): The master port for distributed training backend (str, optional): Backend for ``torch.distributed``, defaults to ``nccl`` seed (int, optional): Specified random seed for every process. Defaults to 1024. verbose (bool, optional): Whether to print logs. Defaults to True. """ try: rank = int(os.environ["OMPI_COMM_WORLD_RANK"]) local_rank = int(os.environ["OMPI_COMM_WORLD_LOCAL_RANK"]) world_size = int(os.environ["OMPI_COMM_WORLD_SIZE"]) except KeyError as e: raise RuntimeError( f"Could not find {e} in the OpenMPI environment, visit https://www.colossalai.org/ for more information on launching with OpenMPI" ) launch( local_rank=local_rank, rank=rank, world_size=world_size, host=host, port=port, backend=backend, seed=seed, verbose=verbose, ) def launch_from_torch(backend: str = "nccl", seed: int = 1024, verbose: bool = True): """A wrapper for colossalai.launch for torchrun or torch.distributed.launch by reading rank and world size from the environment variables set by PyTorch Args: config (Union[str, dict, Config]): Config file or config file path are both acceptable backend (str, optional): Backend for ``torch.distributed``, defaults to ``nccl`` seed (int, optional): Specified random seed for every process. Defaults to 1024. verbose (bool, optional): Whether to print logs. Defaults to True. """ try: rank = int(os.environ["RANK"]) local_rank = int(os.environ["LOCAL_RANK"]) world_size = int(os.environ["WORLD_SIZE"]) host = os.environ["MASTER_ADDR"] port = int(os.environ["MASTER_PORT"]) except KeyError as e: raise RuntimeError( f"Could not find {e} in the torch environment, visit https://www.colossalai.org/ for more information on launching with torch" ) launch( local_rank=local_rank, rank=rank, world_size=world_size, host=host, port=port, backend=backend, seed=seed, verbose=verbose, )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/__init__.py
colossalai/_analyzer/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/envs.py
colossalai/_analyzer/envs.py
from dataclasses import dataclass @dataclass class MeshConfig: TFLOPS: float = 1.9e12 BANDWIDTH = 1.2e9
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/_subclasses/meta_tensor.py
colossalai/_analyzer/_subclasses/meta_tensor.py
import uuid from functools import partial import torch import torch.distributed as dist from torch.types import _device from torch.utils._pytree import tree_map from ._monkey_patch import _AliasATen, _DistCommMethod, _InplaceATen, _MaybeInplaceATen, _TorchOverrideableFactoryMethod __all__ = ["MetaTensor", "MetaTensorMode"] def register_storage(r, data_ptr_fn=None): if isinstance(r, torch.Tensor): if data_ptr_fn is not None: r.data_ptr = data_ptr_fn elif not r.data_ptr(): data_ptr = uuid.uuid1() r.data_ptr = lambda: data_ptr def _normalize_tuple(x): if not isinstance(x, tuple): return (x,) return x # a hack of inplace execution in PyTorch def _assert_alias(func): return func in (_AliasATen + _InplaceATen + _MaybeInplaceATen) # TODO: check if should be this aggressive class MetaTensor(torch.Tensor): """ A wrapping tensor that hacks ``torch.autograd`` without patching more ``torch.ops.aten`` ops. `device` is the device that ``MetaTensor`` is supposed to run on. Meta tensors give you the ability to run PyTorch code without having to actually do computation through tensors allocated on a `meta` device. Because the device is `meta`, meta tensors do not model device propagation. ``MetaTensor`` extends its usage by carrying an additional `device` which tracks devices that would have been used. Reference: https://github.com/pytorch/pytorch/blob/master/torch/_subclasses/fake_tensor.py """ _tensor: torch.Tensor @staticmethod def __new__(cls, elem, device=None, data_ptr_fn=None): requires_grad = elem.requires_grad # Avoid multiple wrapping while isinstance(elem, MetaTensor): device = elem.device if device is None else device elem = elem._tensor # The wrapping tensor (MetaTensor) shouldn't hold any # memory for the class in question, but it should still # advertise the same device as before r = torch.Tensor._make_wrapper_subclass( cls, elem.size(), strides=elem.stride(), storage_offset=elem.storage_offset(), dtype=elem.dtype, layout=elem.layout, device=device or (elem.device if elem.device.type != "meta" else torch.device("cpu")), requires_grad=requires_grad, ) # deceive the frontend for aten selections r._tensor = elem # ...the real tensor is held as an element on the tensor. if not r._tensor.is_meta: val = elem.data_ptr() data_ptr_fn = lambda: val r._tensor = r._tensor.to(torch.device("meta")) # only tensor not on `meta` should be copied to `meta` register_storage(r._tensor, data_ptr_fn) if isinstance(elem, torch.nn.Parameter): r = torch.nn.Parameter(r) return r def __repr__(self): name = "MetaParameter" if getattr(self, "_is_param", False) else "MetaTensor" if self.grad_fn: return f"{name}(..., size={tuple(self.shape)}, device='{self.device}', dtype={self.dtype}, grad_fn={self.grad_fn})" return f"{name}(..., size={tuple(self.shape)}, device='{self.device}', dtype={self.dtype})" @classmethod def __torch_dispatch__(cls, func, types, args=(), kwargs=None): device = None def unwrap(x): nonlocal device if isinstance(x, MetaTensor): device = x.device x = x._tensor elif isinstance(x, torch.Tensor): device = x.device x = x.to(torch.device("meta")) return x args = tree_map(unwrap, args) kwargs = tree_map(unwrap, kwargs) if "device" in kwargs: device = kwargs["device"] kwargs["device"] = torch.device("meta") # run aten for backend=CPU but actually on backend=Meta # here we detect whether or not the execution generates a physical copy # of the input tensor ret = func(*args, **kwargs) if _assert_alias(func): val = args[0].data_ptr() tree_map(partial(register_storage, data_ptr_fn=lambda: val), _normalize_tuple(ret)) # Now, we want to continue propagating this tensor, so we rewrap Tensors in # our custom tensor subclass def wrap(x): return MetaTensor(x, device=device) if isinstance(x, torch.Tensor) else x return tree_map(wrap, ret) def to(self, *args, **kwargs) -> torch.Tensor: """An extension of `torch.Tensor.to()` to MetaTensor Returns: result (MetaTensor): MetaTensor Usage: >>> tensor = MetaTensor(torch.rand(10), device='cuda:100') >>> tensor.to(torch.uint8) MetaTensor(tensor(..., device='meta', size=(10,), dtype=torch.uint8), device='cuda:100') >>> tensor.to(torch.device('cuda:42')) MetaTensor(tensor(..., device='meta', size=(10,)), device='cuda:42') >>> tensor.to('vulkan') MetaTensor(tensor(..., device='meta', size=(10,)), device='vulkan') """ # this imitates c++ function in the way of @overload device = None def replace(x): nonlocal device if isinstance(x, str) or isinstance(x, _device): device = x return torch.device("meta") return x elem = self._tensor.to(*tree_map(replace, args), **tree_map(replace, kwargs)) return MetaTensor(elem, device=device) def cpu(self, *args, **kwargs): if self.device.type == "cpu": return self.to(*args, **kwargs) return self.to(*args, device="cpu", **kwargs) def cuda(self, device=None, non_blocking=False): if device is not None: return self.to(device=device, non_blocking=non_blocking) return self.to(device="cuda:0", non_blocking=non_blocking) def data_ptr(self): return self._tensor.data_ptr() class MetaTensorMode(object): """ A context manager that enables MetaTensor mode. Usage: >>> with MetaTensorMode(): >>> # all torch.xxx and torch.distributed.xxx will be replaced by patched functions >>> # and the actual execution will be on torch.device('meta') >>> a = torch.rand(100000, 100000) >>> b = torch.rand(100000, 100000) >>> c = torch.mm(a, b) """ def __init__(self): self.torch_overrides = {} # override torch.xxx self.dist_overrides = {} # override torch.distributed.xxx def __enter__(self): def _dummy(*args, **kwargs): pass def _new(*args, orig_new=torch.empty, **kwargs): return MetaTensor( orig_new(*args, **{**kwargs, "device": "meta"}), device=kwargs.get("device", torch.device("cpu")) ) for func in _TorchOverrideableFactoryMethod: self.torch_overrides[func] = getattr(torch, func) setattr(torch, func, partial(_new, orig_new=getattr(torch, func))) for func in _DistCommMethod: self.dist_overrides[func] = getattr(dist, func) setattr(dist, func, _dummy) def __exit__(self, exc_type, exc_value, traceback): for func, func_impl in self.torch_overrides.items(): setattr(torch, func, func_impl) for func, func_impl in self.dist_overrides.items(): setattr(dist, func, func_impl)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/_subclasses/_monkey_patch.py
colossalai/_analyzer/_subclasses/_monkey_patch.py
import torch from packaging import version __all__ = [ "_TorchFactoryMethod", "_TorchOverrideableFactoryMethod", "_TorchNonOverrideableFactoryMethod", "_TensorPropertyMethod", "_DistCommMethod", "_AliasATen", "_InplaceATen", "_MaybeInplaceATen", ] _TorchOverrideableFactoryMethod = [ "empty", "eye", "full", "ones", "rand", "randn", "zeros", ] _TorchNonOverrideableFactoryMethod = [ "arange", "finfo", "linspace", "logspace", "randint", "randperm", "tensor", ] _TorchFactoryMethod = _TorchOverrideableFactoryMethod + _TorchNonOverrideableFactoryMethod _TensorPropertyMethod = ["dtype", "shape", "device", "requires_grad", "grad", "grad_fn", "data"] _DistCommMethod = [ "all_gather", "all_reduce", "all_to_all", "broadcast", "gather", "reduce", "reduce_scatter", "scatter", ] if version.parse(torch.__version__) >= version.parse("1.12.0"): aten = torch.ops.aten # TODO: dive deep here # refer to https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/TensorShape.cpp _AliasATen = [ aten.detach.default, aten.detach_.default, aten.t.default, aten.transpose.int, aten.view.default, aten._unsafe_view.default, aten._reshape_alias.default, ] _InplaceATen = [ aten.add_.Tensor, aten.add_.Scalar, aten.sub_.Tensor, aten.sub_.Scalar, aten.mul_.Tensor, aten.mul_.Scalar, aten.div_.Tensor, aten.div_.Scalar, aten.pow_.Tensor, aten.pow_.Scalar, ] # use `MaybeInplace` because they call ``as_strided()`` or ``slice()`` _MaybeInplaceATen = [ aten.diagonal.default, aten.expand.default, aten.select.int, aten.slice.Tensor, aten.split.Tensor, aten.squeeze.default, aten.permute.default, aten.unsqueeze.default, aten.as_strided.default, ] else: _AliasATen = [] _InplaceATen = [] _MaybeInplaceATen = []
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/_subclasses/_meta_registration.py
colossalai/_analyzer/_subclasses/_meta_registration.py
# meta patch from https://github.com/pytorch/pytorch/blob/master/torch/_meta_registrations.py # should be activated for PyTorch version 1.12.0 and below # refer to https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml # for more meta_registrations from typing import List, Optional, Union import torch from packaging import version from torch.utils._pytree import tree_map aten = torch.ops.aten try: meta_lib = torch.library.Library("aten", "IMPL", "Meta") except AttributeError: meta_lib = None meta_table = {} orig_empty = torch.empty orig_empty_strided = torch.empty_strided orig_empty_like = torch.empty_like def new(*args, **kwargs): return orig_empty(*args, **kwargs, device=torch.device("meta")) def new_strided(*args, **kwargs): return orig_empty_strided(*args, **kwargs, device=torch.device("meta")) def new_like(*args, **kwargs): return orig_empty_like(*args, **kwargs, device=torch.device("meta")) def register_meta(op, register_dispatcher=True): def wrapper(f): def add_func(op): meta_table[op] = f if register_dispatcher: name = op.__name__ if op._overloadname != "default" else op.overloadpacket.__name__ try: meta_lib.impl(name, f) except: pass tree_map(add_func, op) return f return wrapper if version.parse(torch.__version__) >= version.parse("1.12.0"): # ============================== Convolutions ====================================== # https://github.com/pytorch/pytorch/pull/79834 @register_meta(aten.convolution.default) def meta_conv( input_tensor: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, stride: List[int], padding: List[int], dilation: List[int], is_transposed: bool, output_padding: List[int], groups: int, ): def _formula(ln: int, p: int, d: int, k: int, s: int) -> int: """ Formula to apply to calculate the length of some dimension of the output See: https://pytorch.org/docs/stable/generated/torch.nn.Conv2d.html Args: ln: length of the dimension p: padding in that dim d: dilation in that dim k: kernel size in that dim s: stride in that dim Returns: The output length """ return (ln + 2 * p - d * (k - 1) - 1) // s + 1 def _formula_transposed(ln: int, p: int, d: int, k: int, s: int, op: int) -> int: """ Formula to apply to calculate the length of some dimension of the output if transposed convolution is used. See: https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose2d.html Args: ln: length of the dimension p: padding in that dim d: dilation in that dim k: kernel size in that dim s: stride in that dim op: output padding in that dim Returns: The output length """ return (ln - 1) * s - 2 * p + d * (k - 1) + op + 1 def calc_conv_nd_return_shape( dims: torch.Size, kernel_size: torch.Size, stride: Union[List[int], int], padding: Union[List[int], int], dilation: Union[List[int], int], output_padding: Optional[Union[List[int], int]] = None, ): ret_shape = [] if isinstance(stride, int): stride = [stride] * len(dims) elif len(stride) == 1: stride = [stride[0]] * len(dims) if isinstance(padding, int): padding = [padding] * len(dims) elif len(padding) == 1: padding = [padding[0]] * len(dims) if isinstance(dilation, int): dilation = [dilation] * len(dims) elif len(dilation) == 1: dilation = [dilation[0]] * len(dims) output_padding_list: Optional[List[int]] = None if output_padding: if isinstance(output_padding, int): output_padding_list = [output_padding] * len(dims) elif len(output_padding) == 1: output_padding_list = [output_padding[0]] * len(dims) else: output_padding_list = output_padding for i in range(len(dims)): # If output_padding is present, we are dealing with a transposed convolution if output_padding_list: ret_shape.append( _formula_transposed( dims[i], padding[i], dilation[i], kernel_size[i], stride[i], output_padding_list[i], ) ) else: ret_shape.append(_formula(dims[i], padding[i], dilation[i], kernel_size[i], stride[i])) return ret_shape def pick_memory_format(): if input_tensor.is_contiguous(memory_format=torch.channels_last): return torch.channels_last elif input_tensor.is_contiguous(memory_format=torch.contiguous_format): return torch.contiguous_format elif input_tensor.is_contiguous(memory_format=torch.preserve_format): return torch.preserve_format kernel_size = weight.shape[2:] dims = input_tensor.shape[2:] if is_transposed: out_channels = groups * weight.shape[1] shape_out = calc_conv_nd_return_shape( dims, kernel_size, stride, padding, dilation, output_padding, ) else: out_channels = weight.shape[0] if weight.shape[1] != input_tensor.shape[1] / groups: raise RuntimeError("Invalid channel dimensions") shape_out = calc_conv_nd_return_shape(dims, kernel_size, stride, padding, dilation) out = input_tensor.new_empty((input_tensor.shape[0], out_channels, *shape_out)) mem_fmt = pick_memory_format() out = out.to(memory_format=mem_fmt) # type: ignore[call-overload] return out @register_meta(aten._convolution.default) def meta__conv( input_tensor: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, stride: List[int], padding: List[int], dilation: List[int], is_transposed: bool, output_padding: List[int], groups: int, *extra_args, ): out = meta_conv(input_tensor, weight, bias, stride, padding, dilation, is_transposed, output_padding, groups) return out @register_meta(aten.convolution_backward.default) def meta_conv_backward( grad_output: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, bias_sizes, stride, padding, dilation, transposed, output_padding, groups, output_mask, ): return new_like(input), new_like(weight), new((bias_sizes)) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/AdaptiveAveragePooling.cpp @register_meta(aten._adaptive_avg_pool2d_backward.default) def meta_adaptive_avg_pool2d_backward( grad_output: torch.Tensor, input: torch.Tensor, ): return new_like(input) # ================================ RNN ============================================= # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/RNN.cpp @register_meta(aten._cudnn_rnn.default) def meta_cuda_rnn( input, weight, weight_stride0, weight_buf, hx, cx, mode, hidden_size, proj_size, num_layers, batch_first, dropout, train, bidirectional, batch_sizes, dropout_state, ): is_input_packed = len(batch_sizes) != 0 if is_input_packed: seq_length = len(batch_sizes) mini_batch = batch_sizes[0] batch_sizes_sum = input.shape[0] else: seq_length = input.shape[1] if batch_first else input.shape[0] mini_batch = input.shape[0] if batch_first else input.shape[1] batch_sizes_sum = -1 num_directions = 2 if bidirectional else 1 out_size = proj_size if proj_size != 0 else hidden_size if is_input_packed: out_shape = [batch_sizes_sum, out_size * num_directions] else: out_shape = ( [mini_batch, seq_length, out_size * num_directions] if batch_first else [seq_length, mini_batch, out_size * num_directions] ) output = input.new_empty(out_shape) cell_shape = [num_layers * num_directions, mini_batch, hidden_size] cy = new(0) if cx is None else cx.new_empty(cell_shape) hy = hx.new_empty([num_layers * num_directions, mini_batch, out_size]) # TODO: Query cudnnGetRNNTrainingReserveSize (expose to python) reserve_shape = 0 if train else 0 reserve = input.new_empty(reserve_shape, dtype=torch.uint8) return output, hy, cy, reserve, weight_buf # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/RNN.cpp @register_meta(aten._cudnn_rnn_backward.default) def meta_cudnn_rnn_backward( input: torch.Tensor, weight: torch.Tensor, weight_stride0: int, hx: torch.Tensor, cx: Optional[torch.Tensor] = None, *args, **kwargs, ): return ( new_like(input), new_like(weight), new_like(hx), new_like(cx) if cx is not None else new(()), ) # (grad_input, grad_weight, grad_hx, grad_cx) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Activation.cpp # ============================== Activations ======================================= _unregistered_ewise = [ aten.relu.default, aten.prelu.default, aten.hardswish.default, aten.hardtanh.default, aten.hardswish_backward.default, aten.hardtanh_backward.default, ] if version.parse(torch.__version__) < version.parse("2.0.0"): _unregistered_ewise += [ aten.prelu_backward.default, ] @register_meta(_unregistered_ewise) def meta_unregistered_ewise(input: torch.Tensor, *args): return new_like(input) # ============================== Normalization ===================================== # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.native_batch_norm.default) def meta_bn(input: torch.Tensor, weight, bias, running_mean, running_var, training, momentum, eps): n_input = input.size(1) return new_like(input), new((n_input)), new((n_input)) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.native_batch_norm_backward.default) def meta_bn_backward( dY: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, running_mean, running_var, save_mean, save_invstd, train, eps, output_mask, ): return new_like(input), new_like(weight), new_like(weight) # (dX, dgamma, dbeta) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.cudnn_batch_norm.default) def meta_cudnn_bn(input: torch.Tensor, weight, bias, running_mean, running_var, training, momentum, eps): n_input = input.size(1) return ( new_like(input), new((n_input)), new((n_input)), new((0), dtype=torch.uint8), ) # (output, running_mean, running_var, reserve) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp # NB: CuDNN only implements the backward algorithm for batchnorm # in training mode (evaluation mode batchnorm has a different algorithm), # which is why this doesn't accept a 'training' parameter. @register_meta(aten.cudnn_batch_norm_backward.default) def meta_cudnn_bn_backward( dY: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, running_mean, running_var, save_mean, save_invstd, eps, reserve, ): return new_like(input), new_like(weight), new_like(weight) # (dX, dgamma, dbeta) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/layer_norm.cpp @register_meta(aten.native_layer_norm.default) def meta_ln(input: torch.Tensor, normalized_shape, weight, bias, eps): bs, n_input = input.size(0), input.size(1) return new_like(input), new((bs, n_input, 1)), new((bs, n_input, 1)) # (output, running_mean, running_var) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/layer_norm.cpp @register_meta(aten.native_layer_norm_backward.default) def meta_ln_backward( dY: torch.Tensor, input: torch.Tensor, normalized_shape, mean, rstd, weight, bias, grad_input_mask ): return new_like(input), new_like(weight), new_like(bias) # (dX, dgamma, dbeta) # ================================== Misc ========================================== # Maybe incorrect # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Im2Col.cpp @register_meta(aten.im2col.default) def meta_im2col(input: torch.Tensor, kernel_size, dilation, padding, stride): return new_like(input) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml @register_meta(aten.roll.default) def meta_roll(input: torch.Tensor, shifts, dims): return input # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Scalar.cpp @register_meta(aten._local_scalar_dense.default) def meta_local_scalar_dense(self: torch.Tensor): return 0 # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/TensorCompare.cpp @register_meta(aten.where.self) def meta_where_self(condition: torch.Tensor, self: torch.Tensor, other: torch.Tensor): result_type = torch.result_type(self, other) return new_like(condition + self + other, dtype=result_type) # ============================== Embedding ========================================= # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Embedding.cpp @register_meta(aten.embedding_dense_backward.default) def meta_embedding_dense_backward( grad_output: torch.Tensor, indices: torch.Tensor, num_weights, padding_idx, scale_grad_by_freq ): return new((num_weights, grad_output.size(-1)), dtype=grad_output.dtype, layout=grad_output.layout) # ============================== Dropout =========================================== # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Dropout.cpp @register_meta(aten.native_dropout.default) def meta_native_dropout_default(input: torch.Tensor, p: float, train: bool = False): # notice that mask is bool return new_like(input), new_like(input, dtype=torch.bool) # (output, mask) # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Dropout.cpp @register_meta(aten.native_dropout_backward.default) def meta_native_dropout_backward_default(grad: torch.Tensor, mask: torch.Tensor, scale: float): return new_like(grad) # (grad_in) if version.parse(torch.__version__) < version.parse("1.13.0"): # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml @register_meta(aten.eye.m_out) def meta_eye(n: int, m: int, out: torch.Tensor): return out @register_meta(aten.index.Tensor) def meta_index_Tensor(self, indices): assert indices, "at least one index must be provided" # aten::index is the internal advanced indexing implementation # checkIndexTensorTypes and expandTensors result: List[Optional[torch.Tensor]] = [] for i, index in enumerate(indices): if index is not None: assert index.dtype in [ torch.long, torch.int8, torch.bool, ], "tensors used as indices must be long, byte or bool tensors" if index.dtype in [torch.int8, torch.bool]: nonzero = index.nonzero() k = len(result) assert k + index.ndim <= self.ndim, f"too many indices for tensor of dimension {self.ndim}" for j in range(index.ndim): assert ( index.shape[j] == self.shape[k + j] ), f"The shape of the mask {index.shape} at index {i} does not match the shape of the indexed tensor {self.shape} at index {k + j}" result.append(nonzero.select(1, j)) else: result.append(index) else: result.append(index) indices = result assert ( len(indices) <= self.ndim ), f"too many indices for tensor of dimension {self.ndim} (got {len(indices)})" # expand_outplace import torch._refs as refs indices = list(refs._maybe_broadcast(*indices)) # add missing null tensors while len(indices) < self.ndim: indices.append(None) # hasContiguousSubspace # true if all non-null tensors are adjacent # See: # https://numpy.org/doc/stable/user/basics.indexing.html#combining-advanced-and-basic-indexing # https://stackoverflow.com/questions/53841497/why-does-numpy-mixed-basic-advanced-indexing-depend-on-slice-adjacency state = 0 has_contiguous_subspace = False for index in indices: if state == 0: if index is not None: state = 1 elif state == 1: if index is None: state = 2 else: if index is not None: break else: has_contiguous_subspace = True # transposeToFront # This is the logic that causes the newly inserted dimensions to show up # at the beginning of the tensor, if they're not contiguous if not has_contiguous_subspace: dims = [] transposed_indices = [] for i, index in enumerate(indices): if index is not None: dims.append(i) transposed_indices.append(index) for i, index in enumerate(indices): if index is None: dims.append(i) transposed_indices.append(index) self = self.permute(dims) indices = transposed_indices # AdvancedIndex::AdvancedIndex # Now we can assume the indices have contiguous subspace # This is simplified from AdvancedIndex which goes to more effort # to put the input and indices in a form so that TensorIterator can # take them. If we write a ref for this, probably that logic should # get implemented before_shape: List[int] = [] after_shape: List[int] = [] replacement_shape: List[int] = [] for dim, index in enumerate(indices): if index is None: if replacement_shape: after_shape.append(self.shape[dim]) else: before_shape.append(self.shape[dim]) else: replacement_shape = list(index.shape) return self.new_empty(before_shape + replacement_shape + after_shape)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/_analyzer/_subclasses/__init__.py
colossalai/_analyzer/_subclasses/__init__.py
from ._meta_registration import * from ._monkey_patch import * from .flop_tensor import flop_count, flop_mapping from .meta_tensor import MetaTensor, MetaTensorMode
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false