repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/__init__.py
colossalai/shardformer/modeling/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/mixtral.py
colossalai/shardformer/modeling/mixtral.py
import inspect import warnings from typing import Callable, List, Optional, Tuple, Union import torch import torch.distributed as dist import torch.nn.functional as F from torch.distributed import ProcessGroup from torch.nn import CrossEntropyLoss from transformers.cache_utils import Cache, DynamicCache from transformers.modeling_attn_mask_utils import ( _prepare_4d_causal_attention_mask, _prepare_4d_causal_attention_mask_for_sdpa, ) from transformers.models.mixtral.modeling_mixtral import ( MixtralModel, MixtralSparseMoeBlock, MoeCausalLMOutputWithPast, MoeModelOutputWithPast, apply_rotary_pos_emb, load_balancing_loss_func, repeat_kv, ) from transformers.utils import is_flash_attn_2_available, logging from colossalai.lazy import LazyInitContext from colossalai.moe._operation import ( DPGradScalerIn, DPGradScalerOut, EPGradScalerIn, EPGradScalerOut, all_to_all_uneven, ) from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.quantization.fp8 import all_reduce_fp8 from colossalai.shardformer.layer._operation import ( all_to_all_comm, gather_forward_split_backward, split_forward_gather_backward, ) from colossalai.shardformer.layer.linear import Linear1D_Col, Linear1D_Row, ParallelModule from colossalai.shardformer.shard import ShardConfig from colossalai.shardformer.shard.utils import set_tensors_to_none from colossalai.tensor.moe_tensor.api import set_moe_tensor_ep_group if is_flash_attn_2_available(): from flash_attn import flash_attn_func from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa _flash_supports_window_size = "window_size" in list(inspect.signature(flash_attn_func).parameters) class EPMixtralSparseMoeBlock(ParallelModule): def __init__(self, *args, **kwargs): raise RuntimeError(f"Please use `from_native_module` to create an instance of {self.__class__.__name__}") def setup_process_groups( self, tp_group: ProcessGroup, moe_dp_group: ProcessGroup, ep_group: ProcessGroup, fp8_communication: bool = False, use_zbv: bool = False, ): assert tp_group is not None assert moe_dp_group is not None assert ep_group is not None # setup ep group self.ep_size = dist.get_world_size(ep_group) self.ep_rank = dist.get_rank(ep_group) self.ep_group = ep_group self.fp8_communication = fp8_communication self.use_zbv = use_zbv if self.num_experts % self.ep_size != 0: raise ValueError("The number of experts must be divisible by the number of expert parallel groups.") self.num_experts_per_ep = self.num_experts // self.ep_size self.expert_start_idx = self.ep_rank * self.num_experts_per_ep held_experts = self.experts[self.expert_start_idx : self.expert_start_idx + self.num_experts_per_ep] set_tensors_to_none(self.experts, exclude=set(held_experts)) # setup moe_dp group self.moe_dp_group = moe_dp_group self.moe_dp_size = moe_dp_group.size() # setup global tp group self.tp_group = tp_group if self.tp_group.size() > 1: for expert in held_experts: expert.w1 = Linear1D_Col.from_native_module( expert.w1, self.tp_group, fp8_communication=self.fp8_communication, use_zbv=self.use_zbv ) expert.w3 = Linear1D_Col.from_native_module( expert.w3, self.tp_group, fp8_communication=self.fp8_communication, use_zbv=self.use_zbv ) expert.w2 = Linear1D_Row.from_native_module( expert.w2, self.tp_group, fp8_communication=self.fp8_communication, use_zbv=self.use_zbv ) for p in self.experts.parameters(): set_moe_tensor_ep_group(p, ep_group) @staticmethod def from_native_module( module: MixtralSparseMoeBlock, tp_group: ProcessGroup, moe_dp_group: ProcessGroup, ep_group: ProcessGroup, *args, **kwargs, ) -> "EPMixtralSparseMoeBlock": # TODO: better init LazyInitContext.materialize(module) module.__class__ = EPMixtralSparseMoeBlock fp8_communication = kwargs.get("fp8_communication", False) module.setup_process_groups(tp_group, moe_dp_group, ep_group, fp8_communication) return module def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: batch_size, sequence_length, hidden_dim = hidden_states.shape hidden_states = hidden_states.view(-1, hidden_dim) # router_logits: (batch * sequence_length, n_experts) router_logits = self.gate(hidden_states) routing_weights = F.softmax(router_logits, dim=1, dtype=torch.float) routing_weights, selected_experts = torch.topk(routing_weights, self.top_k, dim=-1) routing_weights /= routing_weights.sum(dim=-1, keepdim=True) # we cast back to the input dtype routing_weights = routing_weights.to(hidden_states.dtype) selected_experts = selected_experts.t().reshape(-1) selected_experts_idx = selected_experts.argsort() dispatch_states = hidden_states.repeat(self.top_k, 1)[selected_experts_idx] input_split_sizes = selected_experts.bincount(minlength=self.num_experts) output_split_sizes = torch.zeros_like(input_split_sizes) dist.all_to_all_single(output_split_sizes, input_split_sizes, group=self.ep_group) with torch.no_grad(): activate_experts = output_split_sizes[: self.num_experts_per_ep].clone() for i in range(1, self.ep_size): activate_experts += output_split_sizes[i * self.num_experts_per_ep : (i + 1) * self.num_experts_per_ep] activate_experts = (activate_experts > 0).float() if self.fp8_communication: all_reduce_fp8(activate_experts, group=self.moe_dp_group) else: dist.all_reduce(activate_experts, group=self.moe_dp_group) input_split_list = input_split_sizes.view(self.ep_size, self.num_experts_per_ep).sum(dim=-1).tolist() output_split_list = output_split_sizes.view(self.ep_size, self.num_experts_per_ep).sum(dim=-1).tolist() output_states, _ = all_to_all_uneven( dispatch_states, input_split_list, output_split_list, self.ep_group, fp8_communication=self.fp8_communication, ) # compute expert output output_states = EPGradScalerIn.apply(output_states, self.ep_size) if output_states.size(0) > 0: if self.num_experts_per_ep == 1: # no need to split expert = self.experts[self.expert_start_idx] output_states = DPGradScalerIn.apply(output_states, self.moe_dp_size, activate_experts[0]) output_states = expert.act_fn(expert.w1(output_states)) * expert.w3(output_states) output_states = expert.w2(output_states) output_states = DPGradScalerOut.apply(output_states, self.moe_dp_size, activate_experts[0]) else: output_states_splits = output_states.split(output_split_sizes.tolist()) output_states_list = [] for i, split_states in enumerate(output_states_splits): if split_states.size(0) == 0: continue expert = self.experts[self.expert_start_idx + i % self.num_experts_per_ep] split_states = DPGradScalerIn.apply( split_states, self.moe_dp_size, activate_experts[i % self.num_experts_per_ep] ) split_states = expert.act_fn(expert.w1(split_states)) * expert.w3(split_states) split_states = expert.w2(split_states) split_states = DPGradScalerOut.apply( split_states, self.moe_dp_size, activate_experts[i % self.num_experts_per_ep] ) output_states_list.append(split_states) output_states = torch.cat(output_states_list) output_states = EPGradScalerOut.apply(output_states, self.ep_size) dispatch_states, _ = all_to_all_uneven( output_states, output_split_list, input_split_list, self.ep_group, fp8_communication=self.fp8_communication ) recover_experts_idx = torch.empty_like(selected_experts_idx) recover_experts_idx[selected_experts_idx] = torch.arange( selected_experts_idx.size(0), device=selected_experts_idx.device ) dispatch_states = dispatch_states[recover_experts_idx] k_hidden_states = dispatch_states.chunk(self.top_k) output_states = k_hidden_states[0] * routing_weights[:, 0, None] for i in range(1, self.top_k): output_states += k_hidden_states[i] * routing_weights[:, i, None] output_states = output_states.reshape(batch_size, sequence_length, hidden_dim) return output_states, router_logits class MixtralPipelineForwards: """ This class serves as a micro library for forward function substitution of Mixtral models under pipeline setting. """ @staticmethod def mixtral_model_forward( self: MixtralModel, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, output_router_logits: Optional[bool] = None, cache_position: Optional[torch.LongTensor] = None, return_dict: Optional[bool] = None, stage_manager: Optional[PipelineStageManager] = None, hidden_states: Optional[torch.FloatTensor] = None, past_router_logits: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, ): r""" Args: labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): Labels for computing the masked language modeling loss. Indices should either be in `[0, ..., config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. Returns: Example: ```python >>> from transformers import AutoTokenizer, MixtralForCausalLM >>> model = MixtralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> prompt = "Hey, are you conscious? Can you talk to me?" >>> inputs = tokenizer(prompt, return_tensors="pt") >>> # Generate >>> generate_ids = model.generate(inputs.input_ids, max_length=30) >>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." ```""" logger = logging.get_logger(__name__) output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_router_logits = ( output_router_logits if output_router_logits is not None else self.config.output_router_logits ) output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) return_dict = return_dict if return_dict is not None else self.config.use_return_dict # retrieve input_ids and inputs_embeds if stage_manager.is_first_stage(): # retrieve input_ids and inputs_embeds if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time") elif input_ids is not None: batch_size, seq_length = input_ids.shape elif inputs_embeds is not None: batch_size, seq_length, _ = inputs_embeds.shape else: raise ValueError("You have to specify either decoder_input_ids or decoder_inputs_embeds") device = input_ids.device if input_ids is not None else inputs_embeds.device if inputs_embeds is None: inputs_embeds = self.embed_tokens(input_ids) hidden_states = inputs_embeds else: input_shape = hidden_states.shape[:-1] batch_size, seq_length = input_shape device = hidden_states.device seq_length_with_past = seq_length past_key_values_length = 0 # TODO(jianghai): left the recording kv-value tensors as () or None type, this feature may be added in the future. if output_attentions: logger.warning_once("output_attentions=True is not supported for pipeline models at the moment.") output_attentions = False if output_hidden_states: logger.warning_once("output_hidden_states=True is not supported for pipeline models at the moment.") output_hidden_states = False if use_cache: logger.warning_once("use_cache=True is not supported for pipeline models at the moment.") use_cache = False if past_key_values is not None: past_key_values_length = past_key_values[0][0].shape[2] seq_length_with_past = seq_length_with_past + past_key_values_length if position_ids is None: position_ids = torch.arange( past_key_values_length, seq_length + past_key_values_length, dtype=torch.long, device=device, ) position_ids = position_ids.unsqueeze(0).view(-1, seq_length) else: position_ids = position_ids.view(-1, seq_length).long() # embed positions, for the first stage, hidden_states is the input embeddings, # for the other stages, hidden_states is the output of the previous stage if is_flash_attn_2_available(): # 2d mask is passed through the layers attention_mask = attention_mask if (attention_mask is not None and 0 in attention_mask) else None else: # 4d mask is passed through the layers attention_mask = _prepare_4d_causal_attention_mask( attention_mask, (batch_size, seq_length), hidden_states, past_key_values_length, sliding_window=self.config.sliding_window, ) if self.gradient_checkpointing and self.training: if use_cache: logger.warning_once( "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." ) use_cache = False position_embeddings = self.rotary_emb(hidden_states, position_ids) # decoder layers all_hidden_states = () if output_hidden_states else None all_self_attns = () if output_attentions else None all_router_logits = () if output_router_logits else None if cache_position is None: past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0 cache_position = torch.arange( past_seen_tokens, past_seen_tokens + hidden_states.shape[1], device=hidden_states.device ) start_idx, end_idx = stage_index[0], stage_index[1] for idx, decoder_layer in enumerate(self.layers[start_idx:end_idx], start=start_idx): if output_hidden_states: all_hidden_states += (hidden_states,) past_key_value = past_key_values[idx] if past_key_values is not None else None if self.gradient_checkpointing and self.training: def create_custom_forward(module): def custom_forward(*inputs): # None for past_key_value return module(*inputs) return custom_forward layer_outputs = torch.utils.checkpoint.checkpoint( create_custom_forward(decoder_layer), hidden_states, attention_mask, position_ids, None, output_attentions, output_router_logits, use_cache, cache_position, position_embeddings, ) else: layer_outputs = decoder_layer( hidden_states, attention_mask, position_ids, past_key_value, output_attentions, output_router_logits, use_cache, cache_position, position_embeddings, ) hidden_states = layer_outputs[0] if use_cache: next_decoder_cache = (layer_outputs[2 if output_attentions else 1],) if output_attentions: all_self_attns += (layer_outputs[1],) if output_router_logits: all_router_logits += (layer_outputs[-1],) if stage_manager.is_last_stage(): hidden_states = self.norm(hidden_states) # add hidden states from the last decoder layer if output_hidden_states: all_hidden_states += (hidden_states,) next_cache = next_decoder_cache if use_cache else None if output_router_logits and past_router_logits is not None: all_router_logits = past_router_logits + all_router_logits if stage_manager.is_last_stage(): if not return_dict: return tuple( v for v in [hidden_states, next_cache, all_hidden_states, all_self_attns, all_router_logits] if v is not None ) return MoeModelOutputWithPast( last_hidden_state=hidden_states, past_key_values=next_cache, hidden_states=all_hidden_states, attentions=all_self_attns, router_logits=all_router_logits, ) else: if output_router_logits: return { "hidden_states": hidden_states, "past_router_logits": all_router_logits, } else: return { "hidden_states": hidden_states, } @staticmethod def mixtral_for_causal_lm_forward( self, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, labels: Optional[torch.LongTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, output_router_logits: Optional[bool] = None, return_dict: Optional[bool] = None, stage_manager: Optional[PipelineStageManager] = None, hidden_states: Optional[torch.FloatTensor] = None, past_router_logits: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, ): r""" Args: labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): Labels for computing the masked language modeling loss. Indices should either be in `[0, ..., config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. Returns: Example: ```python >>> from transformers import AutoTokenizer, MixtralForCausalLM >>> model = MixtralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> prompt = "Hey, are you conscious? Can you talk to me?" >>> inputs = tokenizer(prompt, return_tensors="pt") >>> # Generate >>> generate_ids = model.generate(inputs.input_ids, max_length=30) >>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." ```""" logger = logging.get_logger(__name__) output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_router_logits = ( output_router_logits if output_router_logits is not None else self.config.output_router_logits ) output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) return_dict = return_dict if return_dict is not None else self.config.use_return_dict # TODO(jianghai): left the recording kv-value tensors as () or None type, this feature may be added in the future. if output_attentions: logger.warning_once("output_attentions=True is not supported for pipeline models at the moment.") output_attentions = False if output_hidden_states: logger.warning_once("output_hidden_states=True is not supported for pipeline models at the moment.") output_hidden_states = False # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn) outputs = MixtralPipelineForwards.mixtral_model_forward( self.model, input_ids=input_ids, attention_mask=attention_mask, position_ids=position_ids, past_key_values=past_key_values, inputs_embeds=inputs_embeds, use_cache=use_cache, output_attentions=output_attentions, output_hidden_states=output_hidden_states, output_router_logits=output_router_logits, return_dict=return_dict, stage_manager=stage_manager, hidden_states=hidden_states, stage_index=stage_index, past_router_logits=past_router_logits, ) past_key_values = None if stage_manager.is_last_stage(): hidden_states = outputs[0] logits = self.lm_head(hidden_states) logits = logits.float() loss = None if labels is not None: # Shift so that tokens < n predict n shift_logits = logits[..., :-1, :].contiguous() shift_labels = labels[..., 1:].contiguous() # Flatten the tokens loss_fct = CrossEntropyLoss() shift_logits = shift_logits.view(-1, self.config.vocab_size) shift_labels = shift_labels.view(-1) # Enable model parallelism shift_labels = shift_labels.to(shift_logits.device) loss = loss_fct(shift_logits, shift_labels) aux_loss = None if output_router_logits: aux_loss = load_balancing_loss_func(outputs[-1], self.num_experts, self.num_experts_per_tok) if labels is not None: loss += self.router_aux_loss_coef * aux_loss if not return_dict: output = (logits,) + outputs[1:] if output_router_logits: output = (aux_loss,) + output return (loss,) + output if loss is not None else output return MoeCausalLMOutputWithPast( loss=loss, aux_loss=aux_loss, logits=logits, past_key_values=None, hidden_states=outputs[0], attentions=None, router_logits=outputs[-1], ) else: out = {} hidden_states = outputs.get("hidden_states") out["hidden_states"] = hidden_states if output_router_logits: out["past_router_logits"] = outputs["past_router_logits"] return out def get_mixtral_flash_attention_forward(shard_config, sp_mode=None, sp_size=None, sp_group=None): logger = logging.get_logger(__name__) from transformers.modeling_utils import ALL_ATTENTION_FUNCTIONS from transformers.models.mixtral.modeling_mixtral import eager_attention_forward def forward( self, hidden_states: torch.Tensor, position_embeddings: Tuple[torch.Tensor, torch.Tensor], attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_value: Optional[Cache] = None, output_attentions: bool = False, cache_position: Optional[torch.LongTensor] = None, use_cache: bool = False, **kwargs, ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Cache]]: if sp_mode is not None: assert sp_mode in ["all_to_all", "split_gather", "ring"], "Invalid sp_mode" assert (sp_size is not None) and ( sp_group is not None ), "Must specify sp_size and sp_group for sequence parallel" if "padding_mask" in kwargs: warnings.warn( "Passing `padding_mask` is deprecated and will be removed in v4.37. Please make sure use `attention_mask` instead.`" ) # overwrite attention_mask with padding_mask attention_mask = kwargs.pop("padding_mask") bsz, q_len, _ = hidden_states.size() # sp: modify sp_len when sequence parallel mode is ring if sp_mode in ["split_gather", "ring"]: q_len *= sp_size query_states = self.q_proj(hidden_states) key_states = self.k_proj(hidden_states) value_states = self.v_proj(hidden_states) # sp: all-to-all comminucation when introducing sequence parallel if sp_mode == "all_to_all": query_states = all_to_all_comm(query_states, sp_group, fp8_communication=shard_config.fp8_communication) key_states = all_to_all_comm(key_states, sp_group, fp8_communication=shard_config.fp8_communication) value_states = all_to_all_comm(value_states, sp_group, fp8_communication=shard_config.fp8_communication) bsz, q_len, _ = query_states.size() query_states = query_states.view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2) key_states = key_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2) value_states = value_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2) kv_seq_len = key_states.shape[-2] if past_key_value is not None: if self.layer_idx is None: raise ValueError( f"The cache structure has changed since version v4.36. If you are using {self.__class__.__name__} " "for auto-regressive decoding with k/v caching, please make sure to initialize the attention class " "with a layer index." ) kv_seq_len += past_key_value.get_usable_length(kv_seq_len, self.layer_idx) # Because the input can be padded, the absolute sequence length depends on the max position id. cos, sin = position_embeddings query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin) if not _flash_supports_window_size: logger.warning_once( "The current flash attention version does not support sliding window attention, for a more memory efficient implementation" " make sure to upgrade flash-attn library." ) if past_key_value is not None: cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position} # Specific to RoPE models key_states, value_states = past_key_value.update(key_states, value_states, self.layer_idx, cache_kwargs) # repeat k/v heads if n_kv_heads < n_heads key_states = repeat_kv(key_states, self.num_key_value_groups) value_states = repeat_kv(value_states, self.num_key_value_groups) 0.0 if not self.training else self.attention_dropout # In PEFT, usually we cast the layer norms in float32 for training stability reasons # therefore the input hidden states gets silently casted in float32. Hence, we need # cast them back in float16 just to be sure everything works as expected. input_dtype = query_states.dtype if input_dtype == torch.float32: if torch.is_autocast_enabled(): target_dtype = torch.get_autocast_gpu_dtype() # Handle the case where the model is quantized elif hasattr(self.config, "_pre_quantization_dtype"): target_dtype = self.config._pre_quantization_dtype else: target_dtype = self.q_proj.weight.dtype logger.warning_once( f"The input hidden states seems to be silently casted in float32, this might be related to" f" the fact you have upcasted embedding or layer norm layers in float32. We will cast back the input in" f" {target_dtype}." ) query_states = query_states.to(target_dtype) key_states = key_states.to(target_dtype) value_states = value_states.to(target_dtype) # Reashape to the expected shape for Flash Attention query_states = query_states.transpose(1, 2) key_states = key_states.transpose(1, 2) value_states = value_states.transpose(1, 2) attention_interface: Callable = eager_attention_forward if self.config._attn_implementation != "eager": if self.config._attn_implementation == "sdpa" and kwargs.get("output_attentions", False): logger.warning_once( "`torch.nn.functional.scaled_dot_product_attention` does not support `output_attentions=True`. Falling back to " 'eager attention. This warning can be removed using the argument `attn_implementation="eager"` when loading the model.' ) else: attention_interface = ALL_ATTENTION_FUNCTIONS[self.config._attn_implementation] attn_output, attn_weights = attention_interface( self, query_states, key_states, value_states, attention_mask, dropout=0.0 if not self.training else self.attention_dropout, scaling=self.scaling, sliding_window=getattr(self.config, "sliding_window", None), # main diff with Llama **kwargs, ) # sp: all-to-all comminucation when introducing sequence parallel if sp_mode == "all_to_all": attn_output = attn_output.reshape(bsz, q_len, self.num_heads * self.head_dim).contiguous() # (1, 8, 128) attn_output = all_to_all_comm( attn_output, sp_group, scatter_dim=1, gather_dim=2, fp8_communication=shard_config.fp8_communication ) # (1, 4, 256) else: attn_output = attn_output.reshape(bsz, q_len, self.hidden_size) attn_output = self.o_proj(attn_output) if not output_attentions: attn_weights = None return attn_output, attn_weights return forward def get_mixtral_flash_attention_model_forward(shard_config, sp_mode=None, sp_size=None, sp_group=None): logger = logging.get_logger(__name__) def forward( self, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None,
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/mistral.py
colossalai/shardformer/modeling/mistral.py
import warnings from typing import List, Optional, Tuple, Union import torch from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.cache_utils import Cache, DynamicCache from transformers.modeling_outputs import ( BaseModelOutputWithPast, CausalLMOutputWithPast, SequenceClassifierOutputWithPast, ) from transformers.models.mistral.modeling_mistral import MistralForCausalLM, MistralModel from transformers.utils import logging from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.shardformer.shard import ShardConfig from ..layer import ColoAttention, dist_cross_entropy logger = logging.get_logger(__name__) class MistralForwards: @staticmethod def mistral_model_forward( self: MistralModel, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, cache_position: Optional[torch.LongTensor] = None, stage_manager: Optional[PipelineStageManager] = None, hidden_states: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, ) -> Union[Tuple, BaseModelOutputWithPast]: if use_cache: logger.warning_once("use_cache=True is not supported for Mistral models at the moment.") use_cache = False output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) # retrieve input_ids and inputs_embeds if stage_manager.is_first_stage(): if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time") elif input_ids is not None: batch_size, seq_length = input_ids.shape elif inputs_embeds is not None: batch_size, seq_length, _ = inputs_embeds.shape else: raise ValueError("You have to specify either input_ids or inputs_embeds") inputs_embeds = self.embed_tokens(input_ids) hidden_states = inputs_embeds else: input_shape = hidden_states.shape[:-1] batch_size, seq_length = input_shape hidden_states.device past_key_values_length = 0 if use_cache and past_key_values is None: past_key_values = DynamicCache() if cache_position is None: past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0 cache_position = torch.arange( past_seen_tokens, past_seen_tokens + hidden_states.shape[1], device=hidden_states.device ) if position_ids is None: position_ids = cache_position.unsqueeze(0) if attention_mask is not None and self.config._attn_implementation == "flash_attention_2" and use_cache: is_padding_right = attention_mask[:, -1].sum().item() != batch_size if is_padding_right: raise ValueError( "You are attempting to perform batched generation with padding_side='right'" " this may lead to unexpected behaviour for Flash Attention version of Mistral. Make sure to " " call `tokenizer.padding_side = 'left'` before tokenizing the input. " ) if shard_config.enable_flash_attention: # in this case, attention_mask is a dict rather than a tensor mask_shape = (batch_size, 1, seq_length, seq_length + past_key_values_length) attention_mask = ColoAttention.prepare_attn_kwargs( mask_shape, hidden_states.dtype, hidden_states.device, q_padding_mask=attention_mask, is_causal=True, ) else: attention_mask = self._update_causal_mask( attention_mask, hidden_states, cache_position, past_key_values, output_attentions ) if self.gradient_checkpointing and self.training: if use_cache: logger.warning_once( "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." ) use_cache = False # decoder layers all_hidden_states = () if output_hidden_states else None all_self_attns = () if output_attentions else None position_embeddings = self.rotary_emb(hidden_states, position_ids) start_idx, end_idx = stage_index[0], stage_index[1] num_ckpt_layers = 0 if self.gradient_checkpointing and self.training: num_ckpt_layers = end_idx - start_idx # TODO: We can replace `gradient_checkpointing_enable` fn and initialize a gradient_checkpointing (List[bool]) for each layer if shard_config.gradient_checkpoint_config is not None: num_ckpt_layers = shard_config.gradient_checkpoint_config.get_num_ckpt_layers( stage=stage_manager.stage, num_stages=stage_manager.num_stages, num_layers=end_idx - start_idx, model_chunk_id=(stage_manager.model_chunk_id if stage_manager.is_interleave else 0), num_model_chunks=stage_manager.num_model_chunks, ) assert num_ckpt_layers <= end_idx - start_idx for idx, decoder_layer in enumerate(self.layers[start_idx:end_idx], start=start_idx): if output_hidden_states: all_hidden_states += (hidden_states,) if idx - start_idx < num_ckpt_layers: layer_outputs = self._gradient_checkpointing_func( decoder_layer.__call__, hidden_states, attention_mask=attention_mask, position_ids=position_ids, past_key_value=past_key_values, output_attentions=output_attentions, use_cache=use_cache, cache_position=cache_position, position_embeddings=position_embeddings, ) else: layer_outputs = decoder_layer( hidden_states, attention_mask=attention_mask, position_ids=position_ids, past_key_value=past_key_values, output_attentions=output_attentions, use_cache=use_cache, cache_position=cache_position, position_embeddings=position_embeddings, ) hidden_states = layer_outputs[0] if use_cache: layer_outputs[2 if output_attentions else 1] if output_attentions: all_self_attns += (layer_outputs[1],) if stage_manager.is_last_stage(): hidden_states = self.norm(hidden_states) # add hidden states from the last decoder layer if output_hidden_states: all_hidden_states += (hidden_states,) next_cache = None if stage_manager.is_last_stage(): return BaseModelOutputWithPast( last_hidden_state=hidden_states, past_key_values=next_cache, hidden_states=all_hidden_states, attentions=all_self_attns, ) else: return {"hidden_states": hidden_states} @staticmethod def mistral_for_causal_lm_forward( self: MistralForCausalLM, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, labels: Optional[torch.LongTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, cache_position: Optional[torch.LongTensor] = None, logits_to_keep: Union[int, torch.Tensor] = 0, stage_manager: Optional[PipelineStageManager] = None, hidden_states: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, ) -> Union[Tuple, CausalLMOutputWithPast]: r""" Args: labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): Labels for computing the masked language modeling loss. Indices should either be in `[0, ..., config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. Returns: Example: ```python >>> from transformers import AutoTokenizer, MistralForCausalLM >>> model = MistralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> prompt = "Hey, are you conscious? Can you talk to me?" >>> inputs = tokenizer(prompt, return_tensors="pt") >>> # Generate >>> generate_ids = model.generate(inputs.input_ids, max_length=30) >>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." ```""" output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn) outputs = MistralForwards.mistral_model_forward( self.model, input_ids=input_ids, attention_mask=attention_mask, position_ids=position_ids, past_key_values=past_key_values, inputs_embeds=inputs_embeds, use_cache=use_cache, output_attentions=output_attentions, output_hidden_states=output_hidden_states, cache_position=cache_position, stage_manager=stage_manager, hidden_states=hidden_states, stage_index=stage_index, shard_config=shard_config, ) past_key_values = None if stage_manager.is_last_stage(): hidden_states = outputs[0] logits = self.lm_head(hidden_states) logits = logits.float() loss = None if labels is not None: loss = dist_cross_entropy(labels, logits, shard_config, self.lm_head.out_features, self.model.dtype) return CausalLMOutputWithPast( loss=loss, logits=logits, past_key_values=outputs.past_key_values, hidden_states=outputs.hidden_states, attentions=outputs.attentions, ) else: hidden_states = outputs.get("hidden_states") return {"hidden_states": hidden_states} @staticmethod def mistral_for_sequence_classification_forward( self, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, labels: Optional[torch.LongTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, stage_manager: Optional[PipelineStageManager] = None, hidden_states: Optional[torch.FloatTensor] = None, stage_index: Optional[List[int]] = None, shard_config: ShardConfig = None, ) -> Union[Tuple, SequenceClassifierOutputWithPast]: r""" labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*): Labels for computing the sequence classification/regression loss. Indices should be in `[0, ..., config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If `config.num_labels > 1` a classification loss is computed (Cross-Entropy). """ transformer_outputs = MistralForwards.mistral_model_forward( self.model, input_ids, attention_mask=attention_mask, position_ids=position_ids, past_key_values=past_key_values, inputs_embeds=inputs_embeds, use_cache=use_cache, output_attentions=output_attentions, output_hidden_states=output_hidden_states, stage_manager=stage_manager, hidden_states=hidden_states, stage_index=stage_index, shard_config=shard_config, ) if input_ids is not None: batch_size = input_ids.shape[0] elif inputs_embeds is not None: batch_size = inputs_embeds.shape[0] else: batch_size = hidden_states.shape[0] if stage_manager.is_last_stage(): hidden_states = transformer_outputs[0] logits = self.score(hidden_states) if self.config.pad_token_id is None and batch_size != 1: raise ValueError("Cannot handle batch sizes > 1 if no padding token is defined.") if self.config.pad_token_id is None: sequence_lengths = -1 else: if input_ids is not None: sequence_lengths = (torch.eq(input_ids, self.config.pad_token_id).int().argmax(-1) - 1).to( logits.device ) else: sequence_lengths = -1 pooled_logits = logits[torch.arange(batch_size, device=logits.device), sequence_lengths] loss = None if labels is not None: labels = labels.to(logits.device) if self.config.problem_type is None: if self.num_labels == 1: self.config.problem_type = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): self.config.problem_type = "single_label_classification" else: self.config.problem_type = "multi_label_classification" if self.config.problem_type == "regression": loss_fct = MSELoss() if self.num_labels == 1: loss = loss_fct(pooled_logits.squeeze(), labels.squeeze()) else: loss = loss_fct(pooled_logits, labels) elif self.config.problem_type == "single_label_classification": loss_fct = CrossEntropyLoss() loss = loss_fct(pooled_logits.view(-1, self.num_labels), labels.view(-1)) elif self.config.problem_type == "multi_label_classification": loss_fct = BCEWithLogitsLoss() loss = loss_fct(pooled_logits, labels) else: hidden_states = transformer_outputs.get("hidden_states") return {"hidden_states": hidden_states} return SequenceClassifierOutputWithPast( loss=loss, logits=pooled_logits, past_key_values=transformer_outputs.past_key_values, hidden_states=transformer_outputs.hidden_states, attentions=transformer_outputs.attentions, ) def get_mistral_model_forward_for_flash_attn(shard_config: ShardConfig): logger = logging.get_logger(__name__) assert shard_config.enable_flash_attention, "Flash Attention is not enabled." def forward( self: MistralModel, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, cache_position: Optional[torch.LongTensor] = None, **flash_attn_kwargs, ) -> Union[Tuple, BaseModelOutputWithPast]: output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) use_cache = use_cache if use_cache is not None else self.config.use_cache # retrieve input_ids and inputs_embeds if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both decoder_input_ids and decoder_inputs_embeds at the same time") elif input_ids is not None: batch_size, seq_length = input_ids.shape elif inputs_embeds is not None: batch_size, seq_length, _ = inputs_embeds.shape else: raise ValueError("You have to specify either decoder_input_ids or decoder_inputs_embeds") if inputs_embeds is None: inputs_embeds = self.embed_tokens(input_ids) if use_cache and past_key_values is None: past_key_values = DynamicCache() if cache_position is None: past_seen_tokens = past_key_values.get_seq_length() if past_key_values is not None else 0 cache_position = torch.arange( past_seen_tokens, past_seen_tokens + inputs_embeds.shape[1], device=inputs_embeds.device ) if position_ids is None: position_ids = cache_position.unsqueeze(0) if attention_mask is not None and self.config._attn_implementation == "flash_attention_2" and use_cache: is_padding_right = attention_mask[:, -1].sum().item() != batch_size if is_padding_right: raise ValueError( "You are attempting to perform batched generation with padding_side='right'" " this may lead to unexpected behaviour for Flash Attention version of Mistral. Make sure to " " call `tokenizer.padding_side = 'left'` before tokenizing the input. " ) if shard_config.enable_flash_attention: # in this case, attention_mask is a dict rather than a tensor mask_shape = (batch_size, 1, seq_length, seq_length) attention_mask = ColoAttention.prepare_attn_kwargs( mask_shape, inputs_embeds.dtype, inputs_embeds.device, q_padding_mask=attention_mask, is_causal=True, ) hidden_states = inputs_embeds position_embeddings = self.rotary_emb(hidden_states, position_ids) if self.gradient_checkpointing and self.training: if use_cache: logger.warning_once( "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." ) use_cache = False # decoder layers all_hidden_states = () if output_hidden_states else None all_self_attns = () if output_attentions else None for decoder_layer in self.layers[: self.config.num_hidden_layers]: if output_hidden_states: all_hidden_states += (hidden_states,) layer_outputs = decoder_layer( hidden_states, attention_mask=attention_mask, position_ids=position_ids, past_key_value=past_key_values, output_attentions=output_attentions, use_cache=use_cache, cache_position=cache_position, position_embeddings=position_embeddings, **flash_attn_kwargs, ) hidden_states = layer_outputs[0] if output_attentions: all_self_attns += (layer_outputs[1],) hidden_states = self.norm(hidden_states) # add hidden states from the last decoder layer if output_hidden_states: all_hidden_states += (hidden_states,) # add hidden states from the last decoder layer if output_hidden_states: all_hidden_states += (hidden_states,) return BaseModelOutputWithPast( last_hidden_state=hidden_states, past_key_values=past_key_values if use_cache else None, hidden_states=all_hidden_states, attentions=all_self_attns, ) return forward def get_mistral_flash_attention_forward(shard_config: ShardConfig): from transformers.models.mistral.modeling_mistral import MistralAttention, apply_rotary_pos_emb, repeat_kv def forward( self: MistralAttention, hidden_states: torch.Tensor, position_embeddings: Tuple[torch.Tensor, torch.Tensor], attention_mask: Optional[torch.Tensor] = None, past_key_value: Optional[Cache] = None, cache_position: Optional[torch.LongTensor] = None, **kwargs, ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]: if "padding_mask" in kwargs: warnings.warn( "Passing `padding_mask` is deprecated and will be removed in v4.37. Please make sure use `attention_mask` instead.`" ) bsz, q_len, _ = hidden_states.size() query_states = self.q_proj(hidden_states) key_states = self.k_proj(hidden_states) value_states = self.v_proj(hidden_states) query_states = query_states.view(bsz, q_len, -1, self.head_dim).transpose(1, 2) key_states = key_states.view(bsz, q_len, -1, self.head_dim).transpose(1, 2) value_states = value_states.view(bsz, q_len, -1, self.head_dim).transpose(1, 2) kv_seq_len = key_states.shape[-2] if past_key_value is not None: if self.layer_idx is None: raise ValueError( f"The cache structure has changed since version v4.36. If you are using {self.__class__.__name__} " "for auto-regressive decoding with k/v caching, please make sure to initialize the attention class " "with a layer index." ) kv_seq_len += past_key_value.get_usable_length(kv_seq_len, self.layer_idx) # cos, sin = self.rotary_emb(value_states, seq_len=kv_seq_len) cos, sin = position_embeddings query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin) if past_key_value is not None: cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position} # Specific to RoPE models key_states, value_states = past_key_value.update(key_states, value_states, self.layer_idx, cache_kwargs) # repeat k/v heads if n_kv_heads < n_heads key_states = repeat_kv(key_states, self.num_key_value_groups) value_states = repeat_kv(value_states, self.num_key_value_groups) assert isinstance(attention_mask, dict), "Flash Attention Error: attention_mask should be a dict." attn_output = ColoAttention.attention(query_states, key_states, value_states, **attention_mask) attn_output = attn_output.transpose(1, 2).contiguous() attn_output = attn_output.reshape(bsz, q_len, -1) attn_output = self.o_proj(attn_output) return attn_output, None return forward def get_lm_forward_with_dist_cross_entropy(shard_config: ShardConfig): from transformers import MistralForCausalLM def forward( self: MistralForCausalLM, input_ids: torch.LongTensor = None, attention_mask: Optional[torch.Tensor] = None, position_ids: Optional[torch.LongTensor] = None, past_key_values: Optional[List[torch.FloatTensor]] = None, inputs_embeds: Optional[torch.FloatTensor] = None, labels: Optional[torch.LongTensor] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = None, output_hidden_states: Optional[bool] = None, return_dict: Optional[bool] = None, ) -> Union[Tuple, CausalLMOutputWithPast]: r""" Args: labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): Labels for computing the masked language modeling loss. Indices should either be in `[0, ..., config.vocab_size]` or -100 (see `input_ids` docstring). Tokens with indices set to `-100` are ignored (masked), the loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. Returns: Example: ```python >>> from transformers import AutoTokenizer, MistralForCausalLM >>> model = MistralForCausalLM.from_pretrained(PATH_TO_CONVERTED_WEIGHTS) >>> tokenizer = AutoTokenizer.from_pretrained(PATH_TO_CONVERTED_TOKENIZER) >>> prompt = "Hey, are you conscious? Can you talk to me?" >>> inputs = tokenizer(prompt, return_tensors="pt") >>> # Generate >>> generate_ids = model.generate(inputs.input_ids, max_length=30) >>> tokenizer.batch_decode(generate_ids, skip_special_tokens=True, clean_up_tokenization_spaces=False)[0] "Hey, are you conscious? Can you talk to me?\nI'm not conscious, but I can talk to you." ```""" output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions output_hidden_states = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) return_dict = return_dict if return_dict is not None else self.config.use_return_dict # decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn) outputs = self.model( input_ids=input_ids, attention_mask=attention_mask, position_ids=position_ids, past_key_values=past_key_values, inputs_embeds=inputs_embeds, use_cache=use_cache, output_attentions=output_attentions, output_hidden_states=output_hidden_states, return_dict=return_dict, ) hidden_states = outputs[0] logits = self.lm_head(hidden_states) logits = logits.float() loss = None if labels is not None: loss = dist_cross_entropy(labels, logits, shard_config, self.lm_head.out_features, self.model.dtype) if not return_dict: output = (logits,) + outputs[1:] return (loss,) + output if loss is not None else output return CausalLMOutputWithPast( loss=loss, logits=logits, past_key_values=outputs.past_key_values, hidden_states=outputs.hidden_states, attentions=outputs.attentions, ) return forward
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/chatglm2_6b/modeling_chatglm.py
colossalai/shardformer/modeling/chatglm2_6b/modeling_chatglm.py
""" The ChatGLM2-6B License 1. Definitions “Licensor” means the ChatGLM2-6B Model Team that distributes its Software. “Software” means the ChatGLM2-6B model parameters made available under this license. 2. License Grant Subject to the terms and conditions of this License, the Licensor hereby grants to you a non-exclusive, worldwide, non-transferable, non-sublicensable, revocable, royalty-free copyright license to use the Software solely for your non-commercial research purposes. The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. 3. Restriction You will not use, copy, modify, merge, publish, distribute, reproduce, or create derivative works of the Software, in whole or in part, for any commercial, military, or illegal purposes. You will not use the Software for any act that may undermine China's national security and national unity, harm the public interest of society, or infringe upon the rights and interests of human beings. 4. Disclaimer THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. 5. Limitation of Liability EXCEPT TO THE EXTENT PROHIBITED BY APPLICABLE LAW, IN NO EVENT AND UNDER NO LEGAL THEORY, WHETHER BASED IN TORT, NEGLIGENCE, CONTRACT, LIABILITY, OR OTHERWISE WILL ANY LICENSOR BE LIABLE TO YOU FOR ANY DIRECT, INDIRECT, SPECIAL, INCIDENTAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES, OR ANY OTHER COMMERCIAL LOSSES, EVEN IF THE LICENSOR HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. 6. Dispute Resolution This license shall be governed and construed in accordance with the laws of People’s Republic of China. Any dispute arising from or in connection with this License shall be submitted to Haidian District People's Court in Beijing. Note that the license is subject to update to a more comprehensive version. For any questions related to the license and copyright, please contact us at glm-130b@googlegroups.com. """ """ PyTorch ChatGLM model. """ import copy import math import sys import warnings from typing import Any, Callable, Dict, List, Optional, Tuple import torch import torch.nn.functional as F import torch.utils.checkpoint from torch import nn from torch.nn import CrossEntropyLoss, LayerNorm from torch.nn.utils import skip_init from transformers.generation.logits_process import LogitsProcessor from transformers.generation.utils import GenerationConfig, LogitsProcessorList, ModelOutput, StoppingCriteriaList from transformers.modeling_outputs import BaseModelOutputWithPast, CausalLMOutputWithPast from transformers.modeling_utils import PreTrainedModel from transformers.utils import logging from .configuration_chatglm import ChatGLMConfig # flags required to enable jit fusion kernels if sys.platform != "darwin": torch._C._jit_set_profiling_mode(False) torch._C._jit_set_profiling_executor(False) torch._C._jit_override_can_fuse_on_cpu(True) torch._C._jit_override_can_fuse_on_gpu(True) logger = logging.get_logger(__name__) _CHECKPOINT_FOR_DOC = "THUDM/ChatGLM2-6B" _CONFIG_FOR_DOC = "ChatGLM6BConfig" CHATGLM_6B_PRETRAINED_MODEL_ARCHIVE_LIST = [ "THUDM/chatglm2-6b", # See all ChatGLM models at https://huggingface.co/models?filter=chatglm ] def default_init(cls, *args, **kwargs): return cls(*args, **kwargs) class InvalidScoreLogitsProcessor(LogitsProcessor): def __call__(self, input_ids: torch.LongTensor, scores: torch.FloatTensor) -> torch.FloatTensor: if torch.isnan(scores).any() or torch.isinf(scores).any(): scores.zero_() scores[..., 5] = 5e4 return scores class PrefixEncoder(torch.nn.Module): """ The torch.nn model to encode the prefix Input shape: (batch-size, prefix-length) Output shape: (batch-size, prefix-length, 2*layers*hidden) """ def __init__(self, config: ChatGLMConfig): super().__init__() self.prefix_projection = config.prefix_projection if self.prefix_projection: # Use a two-layer MLP to encode the prefix kv_size = config.num_layers * config.kv_channels * config.multi_query_group_num * 2 self.embedding = torch.nn.Embedding(config.pre_seq_len, kv_size) self.trans = torch.nn.Sequential( torch.nn.Linear(kv_size, config.hidden_size), torch.nn.Tanh(), torch.nn.Linear(config.hidden_size, kv_size), ) else: self.embedding = torch.nn.Embedding( config.pre_seq_len, config.num_layers * config.kv_channels * config.multi_query_group_num * 2, ) def forward(self, prefix: torch.Tensor): if self.prefix_projection: prefix_tokens = self.embedding(prefix) past_key_values = self.trans(prefix_tokens) else: past_key_values = self.embedding(prefix) return past_key_values def split_tensor_along_last_dim( tensor: torch.Tensor, num_partitions: int, contiguous_split_chunks: bool = False, ) -> List[torch.Tensor]: """Split a tensor along its last dimension. Arguments: tensor: input tensor. num_partitions: number of partitions to split the tensor contiguous_split_chunks: If True, make each chunk contiguous in memory. Returns: A list of Tensors """ # Get the size and dimension. last_dim = tensor.dim() - 1 last_dim_size = tensor.size()[last_dim] // num_partitions # Split. tensor_list = torch.split(tensor, last_dim_size, dim=last_dim) # Note: torch.split does not create contiguous tensors by default. if contiguous_split_chunks: return tuple(chunk.contiguous() for chunk in tensor_list) return tensor_list class RotaryEmbedding(nn.Module): def __init__(self, dim, original_impl=False, device=None, dtype=None): super().__init__() inv_freq = 1.0 / (10000 ** (torch.arange(0, dim, 2, device=device).to(dtype=dtype) / dim)) self.register_buffer("inv_freq", inv_freq) self.dim = dim self.original_impl = original_impl def forward_impl( self, seq_len: int, n_elem: int, dtype: torch.dtype, device: torch.device, base: int = 10000, ): """Enhanced Transformer with Rotary Position Embedding. Derived from: https://github.com/labmlai/annotated_deep_learning_paper_implementations/blob/master/labml_nn/ transformers/rope/__init__.py. MIT License: https://github.com/labmlai/annotated_deep_learning_paper_implementations/blob/master/license. """ # $\Theta = {\theta_i = 10000^{\frac{2(i-1)}{d}}, i \in [1, 2, ..., \frac{d}{2}]}$ theta = 1.0 / (base ** (torch.arange(0, n_elem, 2, dtype=dtype, device=device) / n_elem)) # Create position indexes `[0, 1, ..., seq_len - 1]` seq_idx = torch.arange(seq_len, dtype=dtype, device=device) # Calculate the product of position index and $\theta_i$ idx_theta = torch.outer(seq_idx, theta).float() cache = torch.stack([torch.cos(idx_theta), torch.sin(idx_theta)], dim=-1) # this is to mimic the behavior of complex32, else we will get different results if dtype in (torch.float16, torch.bfloat16, torch.int8): cache = cache.bfloat16() if dtype == torch.bfloat16 else cache.half() return cache def forward(self, max_seq_len, offset=0): return self.forward_impl( max_seq_len, self.dim, dtype=self.inv_freq.dtype, device=self.inv_freq.device, ) @torch.jit.script def apply_rotary_pos_emb(x: torch.Tensor, rope_cache: torch.Tensor) -> torch.Tensor: # x: [sq, b, np, hn] sq, b, np, hn = x.size(0), x.size(1), x.size(2), x.size(3) rot_dim = rope_cache.shape[-2] * 2 x, x_pass = x[..., :rot_dim], x[..., rot_dim:] # truncate to support variable sizes rope_cache = rope_cache[:sq] xshaped = x.reshape(sq, -1, np, rot_dim // 2, 2) rope_cache = rope_cache.view(sq, -1, 1, xshaped.size(3), 2) x_out2 = torch.stack( [ xshaped[..., 0] * rope_cache[..., 0] - xshaped[..., 1] * rope_cache[..., 1], xshaped[..., 1] * rope_cache[..., 0] + xshaped[..., 0] * rope_cache[..., 1], ], -1, ) x_out2 = x_out2.flatten(3) return torch.cat((x_out2, x_pass), dim=-1) class RMSNorm(torch.nn.Module): def __init__(self, normalized_shape, eps=1e-5, device=None, dtype=None, **kwargs): super().__init__() self.elementwise_affine = True self.normalized_shape = normalized_shape self.weight = torch.nn.Parameter(torch.ones(normalized_shape, device=device, dtype=dtype)) self.eps = eps def forward(self, hidden_states: torch.Tensor): input_dtype = hidden_states.dtype variance = hidden_states.to(torch.float32).pow(2).mean(-1, keepdim=True) hidden_states = hidden_states * torch.rsqrt(variance + self.eps) return (self.weight * hidden_states).to(input_dtype) class CoreAttention(torch.nn.Module): def __init__(self, config: ChatGLMConfig, layer_number): super(CoreAttention, self).__init__() self.apply_query_key_layer_scaling = config.apply_query_key_layer_scaling self.attention_softmax_in_fp32 = config.attention_softmax_in_fp32 if self.apply_query_key_layer_scaling: self.attention_softmax_in_fp32 = True self.layer_number = max(1, layer_number) projection_size = config.kv_channels * config.num_attention_heads # Per attention head and per partition values. self.hidden_size_per_partition = projection_size self.hidden_size_per_attention_head = projection_size // config.num_attention_heads self.num_attention_heads_per_partition = config.num_attention_heads coeff = None self.norm_factor = math.sqrt(self.hidden_size_per_attention_head) if self.apply_query_key_layer_scaling: coeff = self.layer_number self.norm_factor *= coeff self.coeff = coeff self.attention_dropout = torch.nn.Dropout(config.attention_dropout) def forward(self, query_layer, key_layer, value_layer, attention_mask): pytorch_major_version = int(torch.__version__.split(".")[0]) if pytorch_major_version >= 2: query_layer, key_layer, value_layer = [k.permute(1, 2, 0, 3) for k in [query_layer, key_layer, value_layer]] if attention_mask is None and query_layer.shape[2] == key_layer.shape[2]: context_layer = torch.nn.functional.scaled_dot_product_attention( query_layer, key_layer, value_layer, is_causal=True ) else: if attention_mask is not None: attention_mask = ~attention_mask context_layer = torch.nn.functional.scaled_dot_product_attention( query_layer, key_layer, value_layer, attention_mask ) context_layer = context_layer.permute(2, 0, 1, 3) new_context_layer_shape = context_layer.size()[:-2] + (self.hidden_size_per_partition,) context_layer = context_layer.reshape(*new_context_layer_shape) else: # Raw attention scores # [b, np, sq, sk] output_size = ( query_layer.size(1), query_layer.size(2), query_layer.size(0), key_layer.size(0), ) # [sq, b, np, hn] -> [sq, b * np, hn] query_layer = query_layer.view(output_size[2], output_size[0] * output_size[1], -1) # [sk, b, np, hn] -> [sk, b * np, hn] key_layer = key_layer.view(output_size[3], output_size[0] * output_size[1], -1) # preallocating input tensor: [b * np, sq, sk] matmul_input_buffer = torch.empty( output_size[0] * output_size[1], output_size[2], output_size[3], dtype=query_layer.dtype, device=query_layer.device, ) # Raw attention scores. [b * np, sq, sk] matmul_result = torch.baddbmm( matmul_input_buffer, query_layer.transpose(0, 1), # [b * np, sq, hn] key_layer.transpose(0, 1).transpose(1, 2), # [b * np, hn, sk] beta=0.0, alpha=(1.0 / self.norm_factor), ) # change view to [b, np, sq, sk] attention_scores = matmul_result.view(*output_size) # =========================== # Attention probs and dropout # =========================== # attention scores and attention mask [b, np, sq, sk] if self.attention_softmax_in_fp32: attention_scores = attention_scores.float() if self.coeff is not None: attention_scores = attention_scores * self.coeff if attention_mask is None and attention_scores.shape[2] == attention_scores.shape[3]: attention_mask = torch.ones( output_size[0], 1, output_size[2], output_size[3], device=attention_scores.device, dtype=torch.bool, ) attention_mask.tril_() attention_mask = ~attention_mask if attention_mask is not None: attention_scores = attention_scores.masked_fill(attention_mask, float("-inf")) attention_probs = F.softmax(attention_scores, dim=-1) attention_probs = attention_probs.type_as(value_layer) # This is actually dropping out entire tokens to attend to, which might # seem a bit unusual, but is taken from the original Transformer paper. attention_probs = self.attention_dropout(attention_probs) # ========================= # Context layer. [sq, b, hp] # ========================= # value_layer -> context layer. # [sk, b, np, hn] --> [b, np, sq, hn] # context layer shape: [b, np, sq, hn] output_size = ( value_layer.size(1), value_layer.size(2), query_layer.size(0), value_layer.size(3), ) # change view [sk, b * np, hn] value_layer = value_layer.view(value_layer.size(0), output_size[0] * output_size[1], -1) # change view [b * np, sq, sk] attention_probs = attention_probs.view(output_size[0] * output_size[1], output_size[2], -1) # matmul: [b * np, sq, hn] context_layer = torch.bmm(attention_probs, value_layer.transpose(0, 1)) # change view [b, np, sq, hn] context_layer = context_layer.view(*output_size) # [b, np, sq, hn] --> [sq, b, np, hn] context_layer = context_layer.permute(2, 0, 1, 3).contiguous() # [sq, b, np, hn] --> [sq, b, hp] new_context_layer_shape = context_layer.size()[:-2] + (self.hidden_size_per_partition,) context_layer = context_layer.view(*new_context_layer_shape) return context_layer class SelfAttention(torch.nn.Module): """Parallel self-attention layer abstract class. Self-attention layer takes input with size [s, b, h] and returns output of the same size. """ def __init__(self, config: ChatGLMConfig, layer_number, device=None): super(SelfAttention, self).__init__() self.layer_number = max(1, layer_number) self.projection_size = config.kv_channels * config.num_attention_heads # Per attention head and per partition values. self.hidden_size_per_attention_head = self.projection_size // config.num_attention_heads self.num_attention_heads_per_partition = config.num_attention_heads self.multi_query_attention = config.multi_query_attention self.qkv_hidden_size = 3 * self.projection_size if self.multi_query_attention: self.num_multi_query_groups_per_partition = config.multi_query_group_num self.qkv_hidden_size = ( self.projection_size + 2 * self.hidden_size_per_attention_head * config.multi_query_group_num ) self.query_key_value = nn.Linear( config.hidden_size, self.qkv_hidden_size, bias=config.add_bias_linear or config.add_qkv_bias, device=device, **_config_to_kwargs(config), ) self.core_attention = CoreAttention(config, self.layer_number) # Output. self.dense = nn.Linear( self.projection_size, config.hidden_size, bias=config.add_bias_linear, device=device, **_config_to_kwargs(config), ) def _allocate_memory(self, inference_max_sequence_len, batch_size, device=None, dtype=None): if self.multi_query_attention: num_attention_heads = self.num_multi_query_groups_per_partition else: num_attention_heads = self.num_attention_heads_per_partition return torch.empty( inference_max_sequence_len, batch_size, num_attention_heads, self.hidden_size_per_attention_head, dtype=dtype, device=device, ) def forward( self, hidden_states, attention_mask, rotary_pos_emb, kv_cache=None, use_cache=True, ): # hidden_states: [sq, b, h] # ================================================= # Pre-allocate memory for key-values for inference. # ================================================= # ===================== # Query, Key, and Value # ===================== # Attention heads [sq, b, h] --> [sq, b, (np * 3 * hn)] mixed_x_layer = self.query_key_value(hidden_states) if self.multi_query_attention: (query_layer, key_layer, value_layer) = mixed_x_layer.split( [ self.num_attention_heads_per_partition * self.hidden_size_per_attention_head, self.num_multi_query_groups_per_partition * self.hidden_size_per_attention_head, self.num_multi_query_groups_per_partition * self.hidden_size_per_attention_head, ], dim=-1, ) query_layer = query_layer.view( query_layer.size()[:-1] + ( self.num_attention_heads_per_partition, self.hidden_size_per_attention_head, ) ) key_layer = key_layer.view( key_layer.size()[:-1] + ( self.num_multi_query_groups_per_partition, self.hidden_size_per_attention_head, ) ) value_layer = value_layer.view( value_layer.size()[:-1] + ( self.num_multi_query_groups_per_partition, self.hidden_size_per_attention_head, ) ) else: new_tensor_shape = mixed_x_layer.size()[:-1] + ( self.num_attention_heads_per_partition, 3 * self.hidden_size_per_attention_head, ) mixed_x_layer = mixed_x_layer.view(*new_tensor_shape) # [sq, b, np, 3 * hn] --> 3 [sq, b, np, hn] (query_layer, key_layer, value_layer) = split_tensor_along_last_dim(mixed_x_layer, 3) # apply relative positional encoding (rotary embedding) if rotary_pos_emb is not None: query_layer = apply_rotary_pos_emb(query_layer, rotary_pos_emb) key_layer = apply_rotary_pos_emb(key_layer, rotary_pos_emb) # adjust key and value for inference if kv_cache is not None: cache_k, cache_v = kv_cache key_layer = torch.cat((cache_k, key_layer), dim=0) value_layer = torch.cat((cache_v, value_layer), dim=0) if use_cache: kv_cache = (key_layer, value_layer) else: kv_cache = None if self.multi_query_attention: key_layer = key_layer.unsqueeze(-2) key_layer = key_layer.expand( -1, -1, -1, self.num_attention_heads_per_partition // self.num_multi_query_groups_per_partition, -1, ) key_layer = key_layer.contiguous().view( key_layer.size()[:2] + ( self.num_attention_heads_per_partition, self.hidden_size_per_attention_head, ) ) value_layer = value_layer.unsqueeze(-2) value_layer = value_layer.expand( -1, -1, -1, self.num_attention_heads_per_partition // self.num_multi_query_groups_per_partition, -1, ) value_layer = value_layer.contiguous().view( value_layer.size()[:2] + ( self.num_attention_heads_per_partition, self.hidden_size_per_attention_head, ) ) # ================================== # core attention computation # ================================== context_layer = self.core_attention(query_layer, key_layer, value_layer, attention_mask) # ================= # Output. [sq, b, h] # ================= output = self.dense(context_layer) return output, kv_cache def _config_to_kwargs(args): common_kwargs = { "dtype": args.torch_dtype, } return common_kwargs class MLP(torch.nn.Module): """MLP. MLP will take the input with h hidden state, project it to 4*h hidden dimension, perform nonlinear transformation, and project the state back into h hidden dimension. """ def __init__(self, config: ChatGLMConfig, device=None): super(MLP, self).__init__() self.add_bias = config.add_bias_linear # Project to 4h. If using swiglu double the output width, see https://arxiv.org/pdf/2002.05202.pdf self.dense_h_to_4h = nn.Linear( config.hidden_size, config.ffn_hidden_size * 2, bias=self.add_bias, device=device, **_config_to_kwargs(config), ) def swiglu(x): x = torch.chunk(x, 2, dim=-1) return F.silu(x[0]) * x[1] self.activation_func = swiglu # Project back to h. self.dense_4h_to_h = nn.Linear( config.ffn_hidden_size, config.hidden_size, bias=self.add_bias, device=device, **_config_to_kwargs(config), ) def forward(self, hidden_states): # [s, b, 4hp] intermediate_parallel = self.dense_h_to_4h(hidden_states) intermediate_parallel = self.activation_func(intermediate_parallel) # [s, b, h] output = self.dense_4h_to_h(intermediate_parallel) return output class GLMBlock(torch.nn.Module): """A single transformer layer. Transformer layer takes input with size [s, b, h] and returns an output of the same size. """ def __init__(self, config: ChatGLMConfig, layer_number, device=None): super(GLMBlock, self).__init__() self.layer_number = layer_number self.apply_residual_connection_post_layernorm = config.apply_residual_connection_post_layernorm self.fp32_residual_connection = config.fp32_residual_connection LayerNormFunc = RMSNorm if config.rmsnorm else LayerNorm # Layernorm on the input data. self.input_layernorm = LayerNormFunc( config.hidden_size, eps=config.layernorm_epsilon, device=device, dtype=config.torch_dtype, ) # Self attention. self.self_attention = SelfAttention(config, layer_number, device=device) self.hidden_dropout = config.hidden_dropout # Layernorm on the attention output self.post_attention_layernorm = LayerNormFunc( config.hidden_size, eps=config.layernorm_epsilon, device=device, dtype=config.torch_dtype, ) # MLP self.mlp = MLP(config, device=device) def forward( self, hidden_states, attention_mask, rotary_pos_emb, kv_cache=None, use_cache=True, ): # hidden_states: [s, b, h] # Layer norm at the beginning of the transformer layer. layernorm_output = self.input_layernorm(hidden_states) # Self attention. attention_output, kv_cache = self.self_attention( layernorm_output, attention_mask, rotary_pos_emb, kv_cache=kv_cache, use_cache=use_cache, ) # Residual connection. if self.apply_residual_connection_post_layernorm: residual = layernorm_output else: residual = hidden_states layernorm_input = torch.nn.functional.dropout(attention_output, p=self.hidden_dropout, training=self.training) layernorm_input = residual + layernorm_input # Layer norm post the self attention. layernorm_output = self.post_attention_layernorm(layernorm_input) # MLP. mlp_output = self.mlp(layernorm_output) # Second residual connection. if self.apply_residual_connection_post_layernorm: residual = layernorm_output else: residual = layernorm_input output = torch.nn.functional.dropout(mlp_output, p=self.hidden_dropout, training=self.training) output = residual + output return output, kv_cache class GLMTransformer(torch.nn.Module): """Transformer class.""" def __init__(self, config: ChatGLMConfig, device=None): super(GLMTransformer, self).__init__() self.fp32_residual_connection = config.fp32_residual_connection self.post_layer_norm = config.post_layer_norm # Number of layers. self.num_layers = config.num_layers # Transformer layers. def build_layer(layer_number): return GLMBlock(config, layer_number, device=device) self.layers = torch.nn.ModuleList([build_layer(i + 1) for i in range(self.num_layers)]) if self.post_layer_norm: LayerNormFunc = RMSNorm if config.rmsnorm else LayerNorm # Final layer norm before output. self.final_layernorm = LayerNormFunc( config.hidden_size, eps=config.layernorm_epsilon, device=device, dtype=config.torch_dtype, ) self.gradient_checkpointing = False def _get_layer(self, layer_number): return self.layers[layer_number] def forward( self, hidden_states, attention_mask, rotary_pos_emb, kv_caches=None, use_cache: Optional[bool] = True, output_hidden_states: Optional[bool] = False, ): if not kv_caches: kv_caches = [None for _ in range(self.num_layers)] presents = () if use_cache else None if self.gradient_checkpointing and self.training: if use_cache: logger.warning_once( "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." ) use_cache = False all_self_attentions = None all_hidden_states = () if output_hidden_states else None for index in range(self.num_layers): if output_hidden_states: all_hidden_states = all_hidden_states + (hidden_states,) layer = self._get_layer(index) if self.gradient_checkpointing and self.training: layer_ret = torch.utils.checkpoint.checkpoint( layer, hidden_states, attention_mask, rotary_pos_emb, kv_caches[index], use_cache, ) else: layer_ret = layer( hidden_states, attention_mask, rotary_pos_emb, kv_cache=kv_caches[index], use_cache=use_cache, ) hidden_states, kv_cache = layer_ret if use_cache: presents = presents + (kv_cache,) if output_hidden_states: all_hidden_states = all_hidden_states + (hidden_states,) # Final layer norm. if self.post_layer_norm: hidden_states = self.final_layernorm(hidden_states) return hidden_states, presents, all_hidden_states, all_self_attentions class ChatGLMPreTrainedModel(PreTrainedModel): """ An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained models. """ is_parallelizable = False supports_gradient_checkpointing = True config_class = ChatGLMConfig base_model_prefix = "transformer" _no_split_modules = ["GLMBlock"] def _init_weights(self, module: nn.Module): """Initialize the weights.""" return def get_masks(self, input_ids, past_key_values, padding_mask=None): batch_size, seq_length = input_ids.shape full_attention_mask = torch.ones(batch_size, seq_length, seq_length, device=input_ids.device) full_attention_mask.tril_() past_length = 0 if past_key_values: past_length = past_key_values[0][0].shape[0] if past_length: full_attention_mask = torch.cat( ( torch.ones(batch_size, seq_length, past_length, device=input_ids.device), full_attention_mask, ), dim=-1, ) if padding_mask is not None: full_attention_mask = full_attention_mask * padding_mask.unsqueeze(1) if not past_length and padding_mask is not None: full_attention_mask -= padding_mask.unsqueeze(-1) - 1 full_attention_mask = (full_attention_mask < 0.5).bool() full_attention_mask.unsqueeze_(1) return full_attention_mask def get_position_ids(self, input_ids, device): batch_size, seq_length = input_ids.shape position_ids = torch.arange(seq_length, dtype=torch.long, device=device).unsqueeze(0).repeat(batch_size, 1) return position_ids def _set_gradient_checkpointing(self, module, value=False): if isinstance(module, GLMTransformer): module.gradient_checkpointing = value class Embedding(torch.nn.Module): """Language model embeddings.""" def __init__(self, config: ChatGLMConfig, device=None): super(Embedding, self).__init__() self.hidden_size = config.hidden_size # Word embeddings (parallel). self.word_embeddings = nn.Embedding( config.padded_vocab_size, self.hidden_size, dtype=config.torch_dtype, device=device, ) self.fp32_residual_connection = config.fp32_residual_connection def forward(self, input_ids): # Embeddings. words_embeddings = self.word_embeddings(input_ids) embeddings = words_embeddings # Data format change to avoid explicit tranposes : [b s h] --> [s b h]. embeddings = embeddings.transpose(0, 1).contiguous() # If the input flag for fp32 residual connection is set, convert for float. if self.fp32_residual_connection: embeddings = embeddings.float() return embeddings
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/chatglm2_6b/configuration_chatglm.py
colossalai/shardformer/modeling/chatglm2_6b/configuration_chatglm.py
from transformers import PretrainedConfig class ChatGLMConfig(PretrainedConfig): model_type = "chatglm" def __init__( self, num_layers=28, padded_vocab_size=65024, hidden_size=4096, ffn_hidden_size=13696, kv_channels=128, num_attention_heads=32, seq_length=2048, hidden_dropout=0.0, attention_dropout=0.0, layernorm_epsilon=1e-5, rmsnorm=True, apply_residual_connection_post_layernorm=False, post_layer_norm=True, add_bias_linear=False, add_qkv_bias=False, bias_dropout_fusion=True, multi_query_attention=False, multi_query_group_num=1, apply_query_key_layer_scaling=True, attention_softmax_in_fp32=True, fp32_residual_connection=False, quantization_bit=0, pre_seq_len=None, prefix_projection=False, **kwargs, ): self.num_layers = num_layers self.vocab_size = padded_vocab_size self.padded_vocab_size = padded_vocab_size self.hidden_size = hidden_size self.ffn_hidden_size = ffn_hidden_size self.kv_channels = kv_channels self.num_attention_heads = num_attention_heads self.seq_length = seq_length self.hidden_dropout = hidden_dropout self.attention_dropout = attention_dropout self.layernorm_epsilon = layernorm_epsilon self.rmsnorm = rmsnorm self.apply_residual_connection_post_layernorm = apply_residual_connection_post_layernorm self.post_layer_norm = post_layer_norm self.add_bias_linear = add_bias_linear self.add_qkv_bias = add_qkv_bias self.bias_dropout_fusion = bias_dropout_fusion self.multi_query_attention = multi_query_attention self.multi_query_group_num = multi_query_group_num self.apply_query_key_layer_scaling = apply_query_key_layer_scaling self.attention_softmax_in_fp32 = attention_softmax_in_fp32 self.fp32_residual_connection = fp32_residual_connection self.quantization_bit = quantization_bit self.pre_seq_len = pre_seq_len self.prefix_projection = prefix_projection super().__init__(**kwargs)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/modeling/chatglm2_6b/__init__.py
colossalai/shardformer/modeling/chatglm2_6b/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/examples/convergence_benchmark.py
colossalai/shardformer/examples/convergence_benchmark.py
import argparse import math from typing import Any, List, Union import evaluate import torch import torch.distributed as dist from data import GLUEDataBuilder from torch import nn from torch.optim import Adam, Optimizer from torch.utils._pytree import tree_map from torch.utils.data import DataLoader from tqdm import tqdm from transformers import BertConfig, BertForSequenceClassification, get_linear_schedule_with_warmup import colossalai from colossalai.cluster import DistCoordinator from colossalai.shardformer import ShardConfig, ShardFormer def to_device(x: Any, device: torch.device) -> Any: def _to(t: Any): if isinstance(t, torch.Tensor): return t.to(device) return t return tree_map(_to, x) def train(args): colossalai.launch_from_torch(seed=42) coordinator = DistCoordinator() # prepare for data and dataset data_builder = GLUEDataBuilder( model_name_or_path=args.pretrain, task_name=args.task, train_batch_size=args.batch_size, eval_batch_size=args.batch_size, ) train_dataloader = data_builder.train_dataloader() test_dataloader = data_builder.test_dataloader() if args.model == "bert": cfg = BertConfig.from_pretrained(args.pretrain, num_labels=data_builder.num_labels) model = BertForSequenceClassification.from_pretrained(args.pretrain, config=cfg) model.to(torch.cuda.current_device()) # if multiple GPUs, shard the model if dist.get_world_size() > 1: tp_group = dist.new_group(backend="nccl") shard_config = ShardConfig( tensor_parallel_process_group=tp_group, enable_tensor_parallelism=True, enable_all_optimization=True ) shard_former = ShardFormer(shard_config=shard_config) model, _ = shard_former.optimize(model) optim = Adam(model.parameters(), lr=args.lr) num_update_steps_per_epoch = len(train_dataloader) // args.accumulation_steps max_steps = math.ceil(args.max_epochs * num_update_steps_per_epoch) lr_scheduler = get_linear_schedule_with_warmup( optim, num_warmup_steps=math.ceil(max_steps * args.warmup_fraction), num_training_steps=max_steps, ) fit( model, optim, lr_scheduler, train_dataloader, args.max_epochs, args.accumulation_steps, args.batch_size, coordinator, ) results = evaluate_model( model, test_dataloader, data_builder.num_labels, args.task, data_builder.eval_splits, coordinator ) if coordinator.is_master(): print(results) if args.target_f1 is not None and "f1" in results: assert results["f1"] >= args.target_f1, f'f1 score {results["f1"]} is lower than target {args.target_f1}' def fit( model: nn.Module, optimizer: Optimizer, scheduler, train_dataloader, max_epochs, accumulation_steps, batch_size, coordinator, ): step_bar = tqdm( range(len(train_dataloader) // accumulation_steps * max_epochs), desc=f"steps", disable=not coordinator.is_master(), ) total_loss = 0 for epoch in range(max_epochs): model.train() for batch_id, batch in enumerate(train_dataloader): batch = to_device(batch, torch.cuda.current_device()) outputs = model(**batch) loss = outputs.loss loss = loss / accumulation_steps loss.backward() total_loss += loss.item() if (batch_id + 1) % accumulation_steps == 0: optimizer.step() scheduler.step() optimizer.zero_grad() step_bar.set_postfix( {"epoch": epoch, "loss": total_loss / batch_size, "lr": scheduler.get_last_lr()[0]} ) total_loss = 0 step_bar.update() # evaluate @torch.no_grad() def evaluate_model( model: nn.Module, test_dataloader: Union[DataLoader, List[DataLoader]], num_labels: int, task_name: str, eval_splits: List[str], coordinator: DistCoordinator, ): metric = evaluate.load("glue", task_name, process_id=coordinator.rank, num_process=coordinator.world_size) model.eval() def evaluate_subset(dataloader: DataLoader): accum_loss = torch.zeros(1, device=torch.cuda.current_device()) for batch in dataloader: batch = to_device(batch, torch.cuda.current_device()) outputs = model(**batch) val_loss, logits = outputs[:2] accum_loss.add_(val_loss) if num_labels > 1: preds = torch.argmax(logits, axis=1) elif num_labels == 1: preds = logits.squeeze() labels = batch["labels"] metric.add_batch(predictions=preds, references=labels) results = metric.compute() if coordinator.is_master(): results["loss"] = accum_loss.item() / (len(dataloader) * dataloader.batch_size) return results if isinstance(test_dataloader, DataLoader): return evaluate_subset(test_dataloader) else: assert len(test_dataloader) == len(eval_splits) final_results = {} for split, sub_loader in zip(eval_splits, test_dataloader): results = evaluate_subset(sub_loader) final_results.update({f"{k}_{split}": v for k, v in results.items()}) return final_results if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("-t", "--task", default="mrpc", help="GLUE task to run") parser.add_argument("--model", type=str, default="bert") parser.add_argument("--pretrain", type=str, default="bert-base-uncased") parser.add_argument("--max_epochs", type=int, default=1) parser.add_argument("--batch_size", type=int, default=4) parser.add_argument("--lr", type=float, default=2.4e-5) parser.add_argument("--fused_layernorm", type=bool, default=False) parser.add_argument("--accumulation_steps", type=int, default=8) parser.add_argument("--warmup_fraction", type=float, default=0.03) parser.add_argument("--target_f1", type=float, default=None) args = parser.parse_args() train(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/examples/performance_benchmark.py
colossalai/shardformer/examples/performance_benchmark.py
""" Shardformer Benchmark """ import torch import torch.distributed as dist import transformers import triton import colossalai from colossalai.shardformer import ShardConfig, ShardFormer def data_gen(batch_size, seq_length): input_ids = torch.randint(0, seq_length, (batch_size, seq_length), dtype=torch.long) attention_mask = torch.ones((batch_size, seq_length), dtype=torch.long) return dict(input_ids=input_ids, attention_mask=attention_mask) def data_gen_for_sequence_classification(batch_size, seq_length): # LM data gen # the `labels` of LM is the token of the output, cause no padding, use `input_ids` as `labels` data = data_gen(batch_size, seq_length) data["labels"] = torch.ones((batch_size), dtype=torch.long) return data MODEL_CONFIG = transformers.LlamaConfig( num_hidden_layers=4, hidden_size=128, intermediate_size=256, num_attention_heads=4, max_position_embeddings=128, num_labels=16, pad_token_id=2, ) BATCH, N_HEADS, N_CTX, D_HEAD = 4, 8, 4096, 64 model_func = lambda: transformers.LlamaForSequenceClassification(MODEL_CONFIG) # vary seq length for fixed head and batch=4 configs = [ triton.testing.Benchmark( x_names=["N_CTX"], x_vals=[2**i for i in range(8, 13)], line_arg="provider", line_vals=["org_model", "shard_model"], line_names=["org_model", "shard_model"], styles=[("red", "-"), ("blue", "-")], ylabel="ms", plot_name=f"lama_for_sequence_classification-batch-{BATCH}", args={"BATCH": BATCH, "dtype": torch.float16, "model_func": model_func}, ) ] def train(model, data): output = model(**data) loss = output.logits.mean() loss.backward() @triton.testing.perf_report(configs) def bench_shardformer(BATCH, N_CTX, provider, model_func, dtype=torch.float32, device="cuda"): warmup = 10 rep = 100 # prepare data data = data_gen_for_sequence_classification(BATCH, N_CTX) data = {k: v.cuda() for k, v in data.items()} model = model_func().to(device) model.train() if provider == "org_model": fn = lambda: train(model, data) ms = triton.testing.do_bench(fn, warmup=warmup, rep=rep) return ms if provider == "shard_model": shard_config = ShardConfig(enable_fused_normalization=True, enable_tensor_parallelism=True) shard_former = ShardFormer(shard_config=shard_config) sharded_model, _ = shard_former.optimize(model) sharded_model = sharded_model.cuda() fn = lambda: train(sharded_model, data) ms = triton.testing.do_bench(fn, warmup=warmup, rep=rep) return ms # start benchmark, command: # torchrun --standalone --nproc_per_node=2 performance_benchmark.py if __name__ == "__main__": colossalai.launch_from_torch() bench_shardformer.run(save_path=".", print_data=dist.get_rank() == 0)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/examples/data.py
colossalai/shardformer/examples/data.py
import datasets from torch.utils.data import DataLoader from transformers import AutoTokenizer, PreTrainedTokenizer from colossalai.booster.plugin.dp_plugin_base import DPPluginBase class GLUEDataBuilder: task_text_field_map = { "cola": ["sentence"], "sst2": ["sentence"], "mrpc": ["sentence1", "sentence2"], "qqp": ["question1", "question2"], "stsb": ["sentence1", "sentence2"], "mnli": ["premise", "hypothesis"], "qnli": ["question", "sentence"], "rte": ["sentence1", "sentence2"], "wnli": ["sentence1", "sentence2"], "ax": ["premise", "hypothesis"], } glue_task_num_labels = { "cola": 2, "sst2": 2, "mrpc": 2, "qqp": 2, "stsb": 1, "mnli": 3, "qnli": 2, "rte": 2, "wnli": 2, "ax": 3, } loader_columns = [ "datasets_idx", "input_ids", "token_type_ids", "attention_mask", "start_positions", "end_positions", "labels", ] def __init__( self, model_name_or_path: str, plugin: DPPluginBase = None, task_name: str = "mrpc", max_seq_length: int = 128, train_batch_size: int = 32, eval_batch_size: int = 32, **kwargs, ): super().__init__() self.model_name_or_path = model_name_or_path self.task_name = task_name self.max_seq_length = max_seq_length self.train_batch_size = train_batch_size self.eval_batch_size = eval_batch_size self.plugin = plugin self.text_fields = self.task_text_field_map[task_name] self.num_labels = self.glue_task_num_labels[task_name] self.tokenizer: PreTrainedTokenizer = AutoTokenizer.from_pretrained(self.model_name_or_path, use_fast=True) self.setup() def setup(self): self.dataset = datasets.load_dataset("glue", self.task_name) for split in self.dataset.keys(): self.dataset[split] = self.dataset[split].map( self.convert_to_features, batched=True, remove_columns=["label"], ) self.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns] self.dataset[split].set_format(type="torch", columns=self.columns) self.eval_splits = [x for x in self.dataset.keys() if "validation" in x] def prepare_data(self): datasets.load_dataset("glue", self.task_name) AutoTokenizer.from_pretrained(self.model_name_or_path, use_fast=True) def train_dataloader(self): if self.plugin == None: return self.native_prepare_dataloader( self.dataset["train"], batch_size=self.train_batch_size, shuffle=True, drop_last=True ) return self.plugin.prepare_dataloader( self.dataset["train"], batch_size=self.train_batch_size, shuffle=True, drop_last=True ) def val_dataloader(self): if self.plugin == None: return self.native_prepare_dataloader(self.dataset["validation"], batch_size=self.eval_batch_size) if len(self.eval_splits) == 1: return self.plugin.prepare_dataloader(self.dataset["validation"], batch_size=self.eval_batch_size) elif len(self.eval_splits) > 1: return [ self.plugin.prepare_dataloader(self.dataset[x], batch_size=self.eval_batch_size) for x in self.eval_splits ] def test_dataloader(self): if self.plugin == None: return self.native_prepare_dataloader(self.dataset["test"], batch_size=self.train_batch_size) if len(self.eval_splits) == 1: return self.plugin.prepare_dataloader(self.dataset["test"], batch_size=self.eval_batch_size) elif len(self.eval_splits) > 1: return [ self.plugin.prepare_dataloader(self.dataset[x], batch_size=self.eval_batch_size) for x in self.eval_splits ] def convert_to_features(self, example_batch): # Either encode single sentence or sentence pairs if len(self.text_fields) > 1: texts_or_text_pairs = list(zip(example_batch[self.text_fields[0]], example_batch[self.text_fields[1]])) else: texts_or_text_pairs = example_batch[self.text_fields[0]] # Tokenize the text/text pairs features = self.tokenizer.batch_encode_plus( texts_or_text_pairs, max_length=self.max_seq_length, padding="max_length", truncation=True ) # Rename label to labels to make it easier to pass to model forward features["labels"] = example_batch["label"] return features def native_prepare_dataloader(self, dataset, batch_size, shuffle=False, drop_last=False, pin_memory=False): return DataLoader( dataset, batch_size=batch_size, sampler=None, shuffle=shuffle, drop_last=drop_last, pin_memory=pin_memory )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/grad_ckpt_config.py
colossalai/shardformer/shard/grad_ckpt_config.py
from dataclasses import dataclass from typing import List, Optional @dataclass class GradientCheckpointConfig: gradient_checkpointing_ratio: float = 0.0 def get_num_ckpt_layers(self, num_layers: int) -> int: return int(self.gradient_checkpointing_ratio * num_layers) @dataclass class PipelineGradientCheckpointConfig(GradientCheckpointConfig): r""" The pipeline gradient config is designed to provide more flexibility for users to control gradient checkpoint in pipeline parallelism. Combined with PipelineStageManager.set_distribution_config, user can fully control the distribution of layers and checkpointed layers in pipeline parallelism. Refer to https://github.com/hpcaitech/ColossalAI/issues/5509 for more details. It provides the following features: 1. `gradient_checkpointing_ratio`: This is used to control gradient checkpointing more precisely, e.g., set 50% of the layers to use gradient checkpointing. 2. Customize # ckpt layers assigned to each stage. This takes precedence over `gradient_checkpointing_ratio`. """ """ Args: gradient_checkpointing_ratio (Optional[float]): The ratio of gradient checkpointing. It can only be used in pipeline parallelism. Defaults to None. num_stages (Optional[int]): Number of stages in the pipeline. Defaults to None. For sanity check. num_model_chunks (Optional[int]): Number of model chunks (1F1B or Interleaved). Defaults to None. For sanity check. num_model_layers (Optional[int]): Number of model layers. Defaults to None. For sanity check. num_ckpt_layers_per_stage (Optional[List[int]]): Number of checkpointed layers for each stage. Defaults to None. Example 1: num_stages = 8 num_layers = 80 num_model_chunks = 1 num_layers_per_stage = [9, 9, 9, 10, 11, 10, 11, 11] num_ckpt_layers_per_stage = [4, 4, 2, 2, 0, 0, 0, 0] Example 2: num_stages = 4 num_layers = 80 num_model_chunks = 2 num_layers_per_stage = [9, 9, 9, 10, 11, 10, 11, 11] # device 0 holds num_layers_per_stage[0] and num_layers_per_stage[4] layers ... """ num_ckpt_layers_per_stage: Optional[List[int]] = None def __post_init__(self): if self._enable_customized_ckpt_layers_per_stage: assert all([num_ckpt_layers >= 0 for num_ckpt_layers in self.num_ckpt_layers_per_stage]) elif self._enable_gradient_checkpointing_ratio: if not (0 <= self.gradient_checkpointing_ratio <= 1): raise ValueError("gradient_checkpointing_ratio should be in 0% to 100%") @property def _enable_gradient_checkpointing_ratio(self) -> bool: return self.gradient_checkpointing_ratio is not None @property def _enable_customized_ckpt_layers_per_stage(self) -> bool: return self.num_ckpt_layers_per_stage is not None def get_num_ckpt_layers( self, stage: int, num_stages: int, num_layers: int, model_chunk_id: int = 0, num_model_chunks: int = 1 ) -> int: if not self._enable_gradient_checkpointing_ratio and not self._enable_customized_ckpt_layers_per_stage: raise RuntimeError("No checkpointed layers information is provided") if self._enable_customized_ckpt_layers_per_stage: assert len(self.num_ckpt_layers_per_stage) == num_stages * num_model_chunks assert stage <= num_stages and model_chunk_id <= num_model_chunks num_ckpt_layers = self.num_ckpt_layers_per_stage[stage + model_chunk_id * num_stages] assert num_ckpt_layers <= num_layers return num_ckpt_layers else: return int(self.gradient_checkpointing_ratio * num_layers)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/sharder.py
colossalai/shardformer/shard/sharder.py
from types import MethodType from typing import Any, Callable, Dict, List, Optional, Set, Union import torch.nn as nn from torch import Tensor from colossalai.lazy import LazyInitContext from .._utils import getattr_, setattr_ from ..policies.auto_policy import get_autopolicy from ..policies.base_policy import Policy, SubModuleReplacementDescription from .shard_config import ShardConfig from .utils import set_tensors_to_none __all__ = ["ModelSharder", "shard_model"] class ModelSharder(object): r""" Shard the original huggingface model according to the policy Args: policy (:class:`Policy`): The policy to shard the model model (:class:`torch.Module`): The model to shard shard_config: The setting of distributed model """ def __init__(self, model: nn.Module, policy: Policy, shard_config: ShardConfig = None) -> None: self.model = model self.shard_config = shard_config self.policy = get_autopolicy(self.model) if policy is None else policy def shard(self) -> List[Dict[int, Tensor]]: r""" Shard the model according to the policy """ self.policy.set_model(self.model) self.policy.set_shard_config(self.shard_config) self._preprocess() # get shared params before release unheld layers, this avoid misjudgment of shared params (None is None) shared_params = self.policy.get_shared_params() held_layers = self._release_unheld_layers() self._replace_module(include=held_layers) self._materialize() self._postprocess() return shared_params def _preprocess(self) -> None: self.model = self.policy.preprocess() def _postprocess(self) -> None: self.model = self.policy.postprocess() def _replace_module(self, include: Optional[Set[nn.Module]] = None) -> None: r""" Replace the module according to the policy, and replace the module one by one Args: model (:class:`torch.nn.Module`): The model to shard """ module_descriptions = self.policy.module_policy() for layer_cls, module_description in module_descriptions.items(): attr_replacement = module_description.attribute_replacement param_replacement = module_description.param_replacement sub_module_replacement = module_description.sub_module_replacement method_replacement = module_description.method_replacement self._recursive_replace_layer( self.model, layer_cls, attr_replacement, param_replacement, method_replacement, sub_module_replacement, include=include, ) def _recursive_replace_layer( self, module: nn.Module, origin_cls: Union[str, nn.Module], attr_replacement: Dict[str, Any], param_replacement: List[Callable], method_replacement: Dict[str, Callable], sub_module_replacement: List[SubModuleReplacementDescription], include: Optional[Set[nn.Module]] = None, ) -> None: r""" Reverse the replace layer operation Args: module (torch.nn.Module): The object of layer to shard origin_cls (Union[str, torch.nn.Module]): The origin layer class or a string of layer class name attr_replacement (Dict[str, Any]): The attribute dict to modify param_replacement (List[Callable]): The function list to get parameter shard information in policy method_replacement (Dict[str, Callable]): Key is the method name, value is the method for replacement sub_module_replacement ((List[SubModuleReplacementDescription]): The function list to get sub module shard information in policy include (Set[nn.Module], optional): The set of modules to keep on current device when pipeline parallel is enabled. Defaults to None """ if (isinstance(origin_cls, str) and origin_cls == module.__class__.__name__) or ( module.__class__ == origin_cls ): if attr_replacement is not None: self._replace_attr(module, attr_replacement) if param_replacement is not None and (include is None or module in include): self._replace_param(module, param_replacement) if method_replacement is not None: self._replace_method(module, method_replacement) if sub_module_replacement is not None: self._replace_sub_module(module, sub_module_replacement, include) for name, child in module.named_children(): self._recursive_replace_layer( child, origin_cls, attr_replacement, param_replacement, method_replacement, sub_module_replacement, include=include, ) def _replace_attr( self, module: nn.Module, attr_replacement: Dict[str, Any], ) -> None: r""" Replace the attribute of the layer Args: module (:class:`torch.nn.Module`): The object of layer to shard attr_replacement (Dict): The attribute dict to modify """ for k, v in attr_replacement.items(): setattr_(module, k, v, ignore=True) def _replace_param( self, module: nn.Module, param_replacement: List[Callable], ) -> None: r""" Replace the parameter of the layer Args: module (:class:`torch.nn.Module`): The object of layer to shard param_replacement (List[Callable]): The function list to get parameter shard information in policy """ for param_func in param_replacement: param_func(module) def _replace_method(self, module: nn.Module, method_replacement: Dict[str, Callable]): for method_name, new_method in method_replacement.items(): # bind the new method to the module bound_method = MethodType(new_method, module) setattr(module, method_name, bound_method) def _replace_sub_module( self, org_layer: nn.Module, sub_module_replacement: List[SubModuleReplacementDescription], include: Optional[Set[nn.Module]] = None, ) -> None: r""" Shard one layer according to the policy, the layer should be the same class as the key in policy's argument_policy return dict Args: org_layer (torch.nn.Module): The origin layer object to shard sub_module_replacement (List[SubModuleReplacementDescription]): The sub module replacement description list include (Set[nn.Module], optional): The set of modules to keep on current device when pipeline parallel is enabled. Defaults to None """ for description in sub_module_replacement: suffix = description.suffix target_module = description.target_module kwargs = {} if description.kwargs is None else description.kwargs assert target_module is not None, "target_module should not be None" native_sub_module = getattr_(org_layer, suffix, ignore=True) # Skip replacement if submodule is not kept by current device when pipeline parallel is enabled. if (include is not None) and (native_sub_module is not None) and (native_sub_module not in include): continue assert not isinstance( native_sub_module, target_module ), f"The module with suffix {suffix} has been replaced, please check the policy" # if it is None and we are allowed to ignore this module # just skip if description.ignore_if_not_exist and native_sub_module is None: continue try: replace_layer = target_module.from_native_module( native_sub_module, process_group=self.shard_config.tensor_parallel_process_group, **kwargs ) except Exception as e: raise RuntimeError( f"Failed to replace {suffix} of type {native_sub_module.__class__.__qualname__}" f" with {target_module.__qualname__} with the exception: {e}. " "Please check your model configuration or sharding policy, you can set up an issue for us to help you as well." ) setattr_(org_layer, suffix, replace_layer) def _get_recursive_held_layers(self, held_layers: Optional[List[nn.Module]]) -> Optional[List[nn.Module]]: def collect_sub_modules(module: nn.Module): if module is None: return recursive_held_layers.append(module) for name, child in module.named_children(): collect_sub_modules(child) recursive_held_layers = [] for module in held_layers: collect_sub_modules(module) return recursive_held_layers def _release_unheld_layers(self) -> Optional[Set[nn.Module]]: r""" Release the unheld layers in the model """ if self.shard_config and self.shard_config.pipeline_stage_manager: held_layers = self.policy.get_held_layers() set_tensors_to_none(self.model, exclude=set(held_layers)) return set(self._get_recursive_held_layers(held_layers)) return None def _materialize(self) -> None: r""" Materialize the model if lazy initialization is used """ LazyInitContext.materialize(self.model)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/utils.py
colossalai/shardformer/shard/utils.py
from typing import Set import torch.nn as nn def set_tensors_to_none(model: nn.Module, exclude: Set[nn.Module] = set()) -> None: """Set all parameters and buffers of model to None Args: model (nn.Module): The model to set """ if model in exclude: return for child in model.children(): set_tensors_to_none(child, exclude=exclude) for n, p in model.named_parameters(recurse=False): setattr(model, n, None) for n, buf in model.named_buffers(recurse=False): setattr(model, n, None)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/__init__.py
colossalai/shardformer/shard/__init__.py
from .grad_ckpt_config import GradientCheckpointConfig, PipelineGradientCheckpointConfig from .shard_config import ShardConfig from .sharder import ModelSharder from .shardformer import ShardFormer __all__ = ["ShardConfig", "ModelSharder", "ShardFormer", "PipelineGradientCheckpointConfig", "GradientCheckpointConfig"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/shard_config.py
colossalai/shardformer/shard/shard_config.py
import warnings from dataclasses import dataclass, field from typing import Any, Dict, Optional import torch.distributed as dist from torch.distributed import ProcessGroup from colossalai.pipeline.stage_manager import PipelineStageManager from .grad_ckpt_config import GradientCheckpointConfig __all__ = ["ShardConfig"] SUPPORT_SP_MODE = ["split_gather", "ring", "all_to_all", "ring_attn"] @dataclass class ShardConfig: r""" The config for sharding the huggingface model Args: tensor_parallel_process_group (Optional[ProcessGroup]): The process group of tensor parallelism, it's necessary when using tensor parallel. Defaults to None, which is the global process group. pipeline_stage_manager (Optional[PipelineStageManager]): If using pipeline parallelism, it's necessary to specify a pipeline stage manager for inter-process communication in pipeline parallelism. Defaults to None, which means not using pipeline parallelism. enable_tensor_parallelism (bool): Whether to use tensor parallelism. Defaults to True. enable_fused_normalization (bool): Whether to use fused layernorm. Defaults to False. enable_flash_attention (bool, optional): Whether to switch on flash attention. Defaults to False. enable_jit_fused (bool, optional): Whether to switch on JIT fused operators. Defaults to False. enable_sequence_parallelism (bool): Whether to turn on sequence parallelism, which partitions non-tensor-parallel regions along the sequence dimension. Defaults to False. gradient_checkpoint_config (Optional[GradientCheckpointConfig]): The gradient checkpoint config. Defaults to None. enable_all_optimization (bool): Whether to turn on all optimization tools including 'fused normalization', 'flash attention', 'JIT fused operators', 'sequence parallelism' and 'sequence overlap'. Defaults to False. fp8_communication (bool, optional): Whether to enable fp8 communication in model parallelism. Defaults to False. parallel_output (bool): For TP: whether to use parallelize cross entropy computation along the feature dim. For SP: set to True to NOT gather the output along the seq dim. """ tensor_parallel_process_group: Optional[ProcessGroup] = None sequence_parallel_process_group: Optional[ProcessGroup] = None pipeline_stage_manager: Optional[PipelineStageManager] = None enable_tensor_parallelism: bool = True enable_all_optimization: bool = False enable_fused_normalization: bool = False enable_flash_attention: bool = False enable_jit_fused: bool = False enable_sequence_parallelism: bool = False sequence_parallelism_mode: str = None parallel_output: bool = True make_vocab_size_divisible_by: int = 64 gradient_checkpoint_config: Optional[GradientCheckpointConfig] = None extra_kwargs: Dict[str, Any] = field(default_factory=dict) # For ring attention sp_axis: Optional[int] = None pg_mesh: Optional[int] = None inner_ring_size: Optional[int] = None # for moe related moe_dp_group: Optional[ProcessGroup] = None ep_group: Optional[ProcessGroup] = None fp8_communication: bool = False # pipeline_parallel_size: int # data_parallel_size: int # tensor_parallel_mode: Literal['1d', '2d', '2.5d', '3d'] @property def tensor_parallel_size(self): return self._tensor_parallel_size @property def sequence_parallel_size(self): return self._sequence_parallel_size @property def expert_parallel_size(self): return self._expert_parallel_size def __post_init__(self): # turn on all optimization if all_optimization is set to True if self.enable_all_optimization: self._turn_on_all_optimization() if self.enable_sequence_parallelism: self.sequence_parallelism_mode = ( "split_gather" if self.sequence_parallelism_mode is None else self.sequence_parallelism_mode ) assert ( self.sequence_parallelism_mode in SUPPORT_SP_MODE ), f"Sequence parallelism mode {self.sequence_parallelism_mode} is not in the supported list {SUPPORT_SP_MODE}" if self.sequence_parallelism_mode in ["split_gather", "ring"]: assert ( self.enable_tensor_parallelism ), f"sequence parallelism mode {self.sequence_parallelism_mode} can only be used when enable_tensor_parallelism is True" else: if self.sequence_parallelism_mode: self.sequence_parallelism_mode = None warnings.warn( f"The sequence_parallelism_mode will be ignored when enable_sequence_parallelism is False" ) # get the tensor parallel size if not self.enable_tensor_parallelism: self._tensor_parallel_size = 1 else: self._tensor_parallel_size = dist.get_world_size(self.tensor_parallel_process_group) # get the sequence parallel size if not self.enable_sequence_parallelism: self._sequence_parallel_size = 1 else: self._sequence_parallel_size = dist.get_world_size(self.sequence_parallel_process_group) self._expert_parallel_size = dist.get_world_size(self.ep_group) if self.ep_group else 1 def _turn_on_all_optimization(self): """ Turn on all optimization. """ # you can add all the optimization flag here try: from apex.normalization import FusedLayerNorm as ApexFusedLayerNorm # noqa apex_avail = True except ImportError: apex_avail = False warnings.warn("You set enable_all_optimization=True, but apex is not installed.") self.enable_fused_normalization = apex_avail self.enable_flash_attention = True self.enable_jit_fused = True # This can cause non-in-place param sharding when used without ZeRO. # It may also slow down training when seq len is small. Plz enable manually. # self.enable_sequence_parallelism = True
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/shardformer/shard/shardformer.py
colossalai/shardformer/shard/shardformer.py
from typing import Dict, List, Tuple import torch.distributed as dist import torch.nn as nn from torch import Tensor from colossalai.cluster import DistCoordinator from ..policies.base_policy import Policy from .shard_config import ShardConfig from .sharder import ModelSharder class ShardFormer: """ Parallelize model based on the given config and policy Example: ```python from colossalai.shardformer import ShardFormer, ShardConfig from transformers import BertForMaskedLM import colossalai import torch colossalai.launch_from_torch() org_model = BertForMaskedLM.from_pretrained('bert-base-uncased') shard_config = ShardConfig() shard_former = ShardFormer(shard_config=shard_config) model, shared_params = shard_former.optimize(org_model) ``` """ def __init__(self, shard_config: ShardConfig): self.is_distributed = dist.is_initialized() if self.is_distributed: self.coordinator = DistCoordinator() else: self.coordinator = None self.shard_config = shard_config def optimize(self, model: nn.Module, policy: Policy = None) -> Tuple[nn.Module, List[Dict[int, Tensor]]]: r""" This method will optimize the model based on the given policy. Args: model (`torch.nn.Model`): the origin huggingface model shard_config (`ShardConfig`): the config for distribute information policy (`Policy`): the custom policy for sharding Returns: the sharded model and the shared parameters """ sharder = ModelSharder(model=model, shard_config=self.shard_config, policy=policy) shared_params = sharder.shard() return model, shared_params
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/context/singleton_meta.py
colossalai/context/singleton_meta.py
import threading class SingletonMeta(type): """ Thread-safe Singleton Meta with double-checked locking. Reference: https://en.wikipedia.org/wiki/Double-checked_locking """ _instances = {} _lock = threading.Lock() def __call__(cls, *args, **kwargs): # First check (without locking) for performance reasons if cls not in cls._instances: # Acquire a lock before proceeding to the second check with cls._lock: # Second check with lock held to ensure thread safety if cls not in cls._instances: instance = super().__call__(*args, **kwargs) cls._instances[cls] = instance else: assert ( len(args) == 0 and len(kwargs) == 0 ), f"{cls.__name__} is a singleton class and an instance has been created." return cls._instances[cls]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/context/config.py
colossalai/context/config.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import inspect import sys from importlib.machinery import SourceFileLoader from pathlib import Path from colossalai.logging import get_dist_logger class Config(dict): """This is a wrapper class for dict objects so that values of which can be accessed as attributes. Args: config (dict): The dict object to be wrapped. """ def __init__(self, config: dict = None): if config is not None: for k, v in config.items(): self._add_item(k, v) def __missing__(self, key): raise KeyError(key) def __getattr__(self, key): try: value = super(Config, self).__getitem__(key) return value except KeyError: raise AttributeError(key) def __setattr__(self, key, value): super(Config, self).__setitem__(key, value) def _add_item(self, key, value): if isinstance(value, dict): self.__setattr__(key, Config(value)) else: self.__setattr__(key, value) def update(self, config): assert isinstance(config, (Config, dict)), "can only update dictionary or Config objects." for k, v in config.items(): self._add_item(k, v) return self @staticmethod def from_file(filename: str): """Reads a python file and constructs a corresponding :class:`Config` object. Args: filename (str): Name of the file to construct the return object. Returns: :class:`Config`: A :class:`Config` object constructed with information in the file. Raises: AssertionError: Raises an AssertionError if the file does not exist, or the file is not .py file """ # check config path if isinstance(filename, str): filepath = Path(filename).absolute() elif isinstance(filename, Path): filepath = filename.absolute() assert filepath.exists(), f"{filename} is not found, please check your configuration path" # check extension extension = filepath.suffix assert extension == ".py", "only .py files are supported" # import the config as module remove_path = False if filepath.parent not in sys.path: sys.path.insert(0, (filepath)) remove_path = True module_name = filepath.stem source_file = SourceFileLoader(fullname=str(module_name), path=str(filepath)) module = source_file.load_module() # load into config config = Config() for k, v in module.__dict__.items(): if k.startswith("__") or inspect.ismodule(v) or inspect.isclass(v): continue else: config._add_item(k, v) logger = get_dist_logger() logger.debug("variables which starts with __, is a module or class declaration are omitted in config file") # remove module del sys.modules[module_name] if remove_path: sys.path.pop(0) return config class ConfigException(Exception): pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/context/__init__.py
colossalai/context/__init__.py
from .config import Config, ConfigException __all__ = [ "Config", "ConfigException", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/_compatibility.py
colossalai/fx/_compatibility.py
from typing import Callable import torch TORCH_MAJOR = int(torch.__version__.split(".")[0]) TORCH_MINOR = int(torch.__version__.split(".")[1]) if TORCH_MAJOR == 1 and TORCH_MINOR < 12: META_COMPATIBILITY = False elif TORCH_MAJOR == 1 and TORCH_MINOR == 12: META_COMPATIBILITY = True elif TORCH_MAJOR == 1 and TORCH_MINOR == 13: META_COMPATIBILITY = True elif TORCH_MAJOR == 2: META_COMPATIBILITY = True def compatibility(is_backward_compatible: bool = False) -> Callable: """A decorator to make a function compatible with different versions of PyTorch. Args: is_backward_compatible (bool, optional): Whether the function is backward compatible. Defaults to False. Returns: Callable: The decorated function """ def decorator(func): if META_COMPATIBILITY: return func else: if is_backward_compatible: return func else: def wrapper(*args, **kwargs): raise RuntimeError(f"Function `{func.__name__}` is not compatible with PyTorch {torch.__version__}") return wrapper return decorator def is_compatible_with_meta() -> bool: """Check the meta compatibility. Normally it should be called before importing some of the `colossalai.fx` modules. If the meta compatibility is not satisfied, the `colossalai.fx` modules will be replaced by its experimental counterparts. Returns: bool: The meta compatibility """ return META_COMPATIBILITY
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/_meta_regist_12.py
colossalai/fx/_meta_regist_12.py
# meta patch from https://github.com/pytorch/pytorch/blob/master/torch/_meta_registrations.py # should be activated for PyTorch version 1.12.0 and below # refer to https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml # for more meta_registrations from typing import List, Optional, Union import torch from torch.utils._pytree import tree_map aten = torch.ops.aten meta_lib = torch.library.Library("aten", "IMPL", "Meta") meta_table = {} def register_meta(op, register_dispatcher=True): def wrapper(f): def add_func(op): meta_table[op] = f if register_dispatcher: name = op.__name__ if op._overloadname != "default" else op.overloadpacket.__name__ try: meta_lib.impl(name, f) except: pass tree_map(add_func, op) return f return wrapper # ============================== Convolutions ====================================== # https://github.com/pytorch/pytorch/pull/79834 @register_meta(aten.convolution.default) def meta_conv( input_tensor: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, stride: List[int], padding: List[int], dilation: List[int], is_transposed: bool, output_padding: List[int], groups: int, ): def _formula(ln: int, p: int, d: int, k: int, s: int) -> int: """ Formula to apply to calculate the length of some dimension of the output See: https://pytorch.org/docs/stable/generated/torch.nn.Conv2d.html Args: ln: length of the dimension p: padding in that dim d: dilation in that dim k: kernel size in that dim s: stride in that dim Returns: The output length """ return (ln + 2 * p - d * (k - 1) - 1) // s + 1 def _formula_transposed(ln: int, p: int, d: int, k: int, s: int, op: int) -> int: """ Formula to apply to calculate the length of some dimension of the output if transposed convolution is used. See: https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose2d.html Args: ln: length of the dimension p: padding in that dim d: dilation in that dim k: kernel size in that dim s: stride in that dim op: output padding in that dim Returns: The output length """ return (ln - 1) * s - 2 * p + d * (k - 1) + op + 1 def calc_conv_nd_return_shape( dims: torch.Size, kernel_size: torch.Size, stride: Union[List[int], int], padding: Union[List[int], int], dilation: Union[List[int], int], output_padding: Optional[Union[List[int], int]] = None, ): ret_shape = [] if isinstance(stride, int): stride = [stride] * len(dims) elif len(stride) == 1: stride = [stride[0]] * len(dims) if isinstance(padding, int): padding = [padding] * len(dims) elif len(padding) == 1: padding = [padding[0]] * len(dims) if isinstance(dilation, int): dilation = [dilation] * len(dims) elif len(dilation) == 1: dilation = [dilation[0]] * len(dims) output_padding_list: Optional[List[int]] = None if output_padding: if isinstance(output_padding, int): output_padding_list = [output_padding] * len(dims) elif len(output_padding) == 1: output_padding_list = [output_padding[0]] * len(dims) else: output_padding_list = output_padding for i in range(len(dims)): # If output_padding is present, we are dealing with a transposed convolution if output_padding_list: ret_shape.append( _formula_transposed( dims[i], padding[i], dilation[i], kernel_size[i], stride[i], output_padding_list[i], ) ) else: ret_shape.append(_formula(dims[i], padding[i], dilation[i], kernel_size[i], stride[i])) return ret_shape def pick_memory_format(): if input_tensor.is_contiguous(memory_format=torch.channels_last): return torch.channels_last elif input_tensor.is_contiguous(memory_format=torch.contiguous_format): return torch.contiguous_format elif input_tensor.is_contiguous(memory_format=torch.preserve_format): return torch.preserve_format kernel_size = weight.shape[2:] dims = input_tensor.shape[2:] if is_transposed: out_channels = groups * weight.shape[1] shape_out = calc_conv_nd_return_shape( dims, kernel_size, stride, padding, dilation, output_padding, ) else: out_channels = weight.shape[0] if weight.shape[1] != input_tensor.shape[1] / groups: raise RuntimeError("Invalid channel dimensions") shape_out = calc_conv_nd_return_shape(dims, kernel_size, stride, padding, dilation) out = input_tensor.new_empty((input_tensor.shape[0], out_channels, *shape_out)) mem_fmt = pick_memory_format() out = out.to(memory_format=mem_fmt) # type: ignore[call-overload] return out @register_meta(aten._convolution.default) def meta_conv_1( input_tensor: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor, stride: List[int], padding: List[int], dilation: List[int], is_transposed: bool, output_padding: List[int], groups: int, *extra_args, ): out = meta_conv(input_tensor, weight, bias, stride, padding, dilation, is_transposed, output_padding, groups) return out @register_meta(aten.convolution_backward.default) def meta_conv_backward( grad_output: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, bias_sizes, stride, padding, dilation, transposed, output_padding, groups, output_mask, ): return torch.empty_like(input), torch.empty_like(weight), torch.empty((bias_sizes), device="meta") # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/AdaptiveAveragePooling.cpp @register_meta(aten._adaptive_avg_pool2d_backward.default) def meta_adaptive_avg_pool2d_backward( grad_output: torch.Tensor, input: torch.Tensor, ): grad_input = torch.empty_like(input) return grad_input # ================================ RNN ============================================= # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/RNN.cpp @register_meta(aten._cudnn_rnn.default) def meta_cuda_rnn( input, weight, weight_stride0, weight_buf, hx, cx, mode, hidden_size, proj_size, num_layers, batch_first, dropout, train, bidirectional, batch_sizes, dropout_state, ): is_input_packed = len(batch_sizes) != 0 if is_input_packed: seq_length = len(batch_sizes) mini_batch = batch_sizes[0] batch_sizes_sum = input.shape[0] else: seq_length = input.shape[1] if batch_first else input.shape[0] mini_batch = input.shape[0] if batch_first else input.shape[1] batch_sizes_sum = -1 num_directions = 2 if bidirectional else 1 out_size = proj_size if proj_size != 0 else hidden_size if is_input_packed: out_shape = [batch_sizes_sum, out_size * num_directions] else: out_shape = ( [mini_batch, seq_length, out_size * num_directions] if batch_first else [seq_length, mini_batch, out_size * num_directions] ) output = input.new_empty(out_shape) cell_shape = [num_layers * num_directions, mini_batch, hidden_size] cy = torch.empty(0) if cx is None else cx.new_empty(cell_shape) hy = hx.new_empty([num_layers * num_directions, mini_batch, out_size]) # TODO: Query cudnnGetRNNTrainingReserveSize (expose to python) reserve_shape = 0 if train else 0 reserve = input.new_empty(reserve_shape, dtype=torch.uint8) return output, hy, cy, reserve, weight_buf # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/RNN.cpp @register_meta(aten._cudnn_rnn_backward.default) def meta_cudnn_rnn_backward( input: torch.Tensor, weight: torch.Tensor, weight_stride0: int, hx: torch.Tensor, cx: Optional[torch.Tensor] = None, *args, **kwargs, ): print(input, weight, hx, cx) grad_input = torch.empty_like(input) grad_weight = torch.empty_like(weight) grad_hx = torch.empty_like(hx) grad_cx = torch.empty_like(cx) if cx is not None else torch.empty((), device="meta") return grad_input, grad_weight, grad_hx, grad_cx # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Activation.cpp # ============================== Activations ======================================= @register_meta(aten.relu.default) def meta_relu(input: torch.Tensor): return torch.empty_like(input) @register_meta(aten.prelu.default) def meta_prelu(input: torch.Tensor, weight: torch.Tensor): return torch.empty_like(input) @register_meta(aten.hardswish.default) def meta_hardswish(input: torch.Tensor): return torch.empty_like(input) @register_meta(aten.hardtanh.default) def meta_hardtanh(input: torch.Tensor, min, max): return torch.empty_like(input) @register_meta(aten.hardswish_backward.default) def meta_hardswish_backward(grad_out: torch.Tensor, input: torch.Tensor): grad_in = torch.empty_like(input) return grad_in @register_meta(aten.hardtanh_backward.default) def meta_hardtanh_backward(grad_out: torch.Tensor, input: torch.Tensor, min_val: int, max_val: int): grad_in = torch.empty_like(input) return grad_in # ============================== Normalization ===================================== # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.native_batch_norm.default) def meta_bn(input: torch.Tensor, weight, bias, running_mean, running_var, training, momentum, eps): n_input = input.size(1) output = torch.empty_like(input) running_mean = torch.empty((n_input), device="meta") running_var = torch.empty((n_input), device="meta") return output, running_mean, running_var # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.native_batch_norm_backward.default) def meta_bn_backward( dY: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, running_mean, running_var, save_mean, save_invstd, train, eps, output_mask, ): dX = torch.empty_like(input) dgamma = torch.empty_like(weight) dbeta = torch.empty_like(weight) return dX, dgamma, dbeta # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp @register_meta(aten.cudnn_batch_norm.default) def meta_cudnn_bn(input: torch.Tensor, weight, bias, running_mean, running_var, training, momentum, eps): n_input = input.size(1) output = torch.empty_like(input) running_mean = torch.empty((n_input), device="meta") running_var = torch.empty((n_input), device="meta") reserve = torch.empty((0), dtype=torch.uint8, device="meta") return output, running_mean, running_var, reserve # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/cudnn/BatchNorm.cpp # NB: CuDNN only implements the backward algorithm for batchnorm # in training mode (evaluation mode batchnorm has a different algorithm), # which is why this doesn't accept a 'training' parameter. @register_meta(aten.cudnn_batch_norm_backward.default) def meta_cudnn_bn_backward( dY: torch.Tensor, input: torch.Tensor, weight: torch.Tensor, running_mean, running_var, save_mean, save_invstd, eps, reserve, ): dX = torch.empty_like(input) dgamma = torch.empty_like(weight) dbeta = torch.empty_like(weight) return dX, dgamma, dbeta # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/layer_norm.cpp @register_meta(aten.native_layer_norm.default) def meta_ln(input: torch.Tensor, normalized_shape, weight, bias, eps): bs = input.size(0) n_input = input.size(1) output = torch.empty_like(input) running_mean = torch.empty((bs, n_input, 1), device="meta") running_var = torch.empty((bs, n_input, 1), device="meta") return output, running_mean, running_var # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/layer_norm.cpp @register_meta(aten.native_layer_norm_backward.default) def meta_ln_backward( dY: torch.Tensor, input: torch.Tensor, normalized_shape, mean, rstd, weight, bias, grad_input_mask ): dX = torch.empty_like(input) dgamma = torch.empty_like(weight) dbeta = torch.empty_like(bias) return dX, dgamma, dbeta # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/group_norm.cpp @register_meta(aten.native_group_norm_backward.default) def meta_gn_backward(dY: torch.Tensor, input: torch.Tensor, mean, rstd, gamma, N, C, HxW, group, grad_input_mask): dX = torch.empty_like(input) dgamma = torch.empty_like(gamma) dbeta = torch.empty_like(gamma) return dX, dgamma, dbeta # ================================== Misc ========================================== # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/native_functions.yaml @register_meta(aten.roll.default) def meta_roll(input: torch.Tensor, shifts, dims): return input # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Scalar.cpp @register_meta(aten._local_scalar_dense.default) def meta_local_scalar_dense(self: torch.Tensor): return 0 # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/TensorCompare.cpp @register_meta(aten.where.self) def meta_where_self(condition: torch.Tensor, self: torch.Tensor, other: torch.Tensor): result_type = torch.result_type(self, other) return torch.empty_like(condition + self + other, dtype=result_type) @register_meta(aten.index.Tensor) def meta_index_Tensor(self, indices): assert indices, "at least one index must be provided" # aten::index is the internal advanced indexing implementation # checkIndexTensorTypes and expandTensors result: List[Optional[torch.Tensor]] = [] for i, index in enumerate(indices): if index is not None: assert index.dtype in [ torch.long, torch.int8, torch.bool, ], "tensors used as indices must be long, byte or bool tensors" if index.dtype in [torch.int8, torch.bool]: nonzero = index.nonzero() k = len(result) assert k + index.ndim <= self.ndim, f"too many indices for tensor of dimension {self.ndim}" for j in range(index.ndim): assert ( index.shape[j] == self.shape[k + j] ), f"The shape of the mask {index.shape} at index {i} does not match the shape of the indexed tensor {self.shape} at index {k + j}" result.append(nonzero.select(1, j)) else: result.append(index) else: result.append(index) indices = result assert len(indices) <= self.ndim, f"too many indices for tensor of dimension {self.ndim} (got {len(indices)})" # expand_outplace import torch._refs as refs indices = list(refs._maybe_broadcast(*indices)) # add missing null tensors while len(indices) < self.ndim: indices.append(None) # hasContiguousSubspace # true if all non-null tensors are adjacent # See: # https://numpy.org/doc/stable/user/basics.indexing.html#combining-advanced-and-basic-indexing # https://stackoverflow.com/questions/53841497/why-does-numpy-mixed-basic-advanced-indexing-depend-on-slice-adjacency state = 0 has_contiguous_subspace = False for index in indices: if state == 0: if index is not None: state = 1 elif state == 1: if index is None: state = 2 else: if index is not None: break else: has_contiguous_subspace = True # transposeToFront # This is the logic that causes the newly inserted dimensions to show up # at the beginning of the tensor, if they're not contiguous if not has_contiguous_subspace: dims = [] transposed_indices = [] for i, index in enumerate(indices): if index is not None: dims.append(i) transposed_indices.append(index) for i, index in enumerate(indices): if index is None: dims.append(i) transposed_indices.append(index) self = self.permute(dims) indices = transposed_indices # AdvancedIndex::AdvancedIndex # Now we can assume the indices have contiguous subspace # This is simplified from AdvancedIndex which goes to more effort # to put the input and indices in a form so that TensorIterator can # take them. If we write a ref for this, probably that logic should # get implemented before_shape: List[int] = [] after_shape: List[int] = [] replacement_shape: List[int] = [] for dim, index in enumerate(indices): if index is None: if replacement_shape: after_shape.append(self.shape[dim]) else: before_shape.append(self.shape[dim]) else: replacement_shape = list(index.shape) return self.new_empty(before_shape + replacement_shape + after_shape) # ============================== Embedding ========================================= # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Embedding.cpp @register_meta(aten.embedding_dense_backward.default) def meta_embedding_dense_backward( grad_output: torch.Tensor, indices: torch.Tensor, num_weights, padding_idx, scale_grad_by_freq ): return torch.empty( (num_weights, grad_output.size(-1)), dtype=grad_output.dtype, device=grad_output.device, layout=grad_output.layout, ) # ============================== Dropout =========================================== # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Dropout.cpp @register_meta(aten.native_dropout.default) def meta_native_dropout_default(input: torch.Tensor, p: float, train: bool = False): # notice that mask is bool output = torch.empty_like(input) mask = torch.empty_like(input, dtype=torch.bool) return output, mask # https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/native/Dropout.cpp @register_meta(aten.native_dropout_backward.default) def meta_native_dropout_backward_default(grad: torch.Tensor, mask: torch.Tensor, scale: float): return torch.empty_like(grad)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/_meta_regist_13.py
colossalai/fx/_meta_regist_13.py
import torch from torch._meta_registrations import register_meta from torch._prims_common import check aten = torch.ops.aten # since we fix the torch version to 1.13.1, we have to add unimplemented meta ops # all these functions are from here https://github.com/pytorch/pytorch/blob/master/torch/_meta_registrations.py @register_meta([aten.convolution_backward.default]) def meta_convolution_backward( grad_output_, input_, weight_, bias_sizes_opt, stride, padding, dilation, transposed, output_padding, groups, output_mask, ): # High level logic taken from slow_conv3d_backward_cpu which should # be representative of all convolution_backward impls backend_grad_input = None backend_grad_weight = None backend_grad_bias = None if output_mask[0]: backend_grad_input = grad_output_.new_empty(input_.size()) if output_mask[1]: backend_grad_weight = grad_output_.new_empty(weight_.size()) if output_mask[2]: backend_grad_bias = grad_output_.new_empty(bias_sizes_opt) return (backend_grad_input, backend_grad_weight, backend_grad_bias) @register_meta(aten._adaptive_avg_pool2d_backward.default) def meta__adaptive_avg_pool2d_backward(grad_out, self): ndim = grad_out.ndim for i in range(1, ndim): check( grad_out.size(i) > 0, lambda: f"adaptive_avg_pool2d_backward(): Expected grad_output to have non-zero \ size for non-batch dimensions, {grad_out.shape} with dimension {i} being empty", ) check( ndim == 3 or ndim == 4, lambda: f"adaptive_avg_pool2d_backward(): Expected 3D or 4D tensor, but got {self.shape}", ) check( self.dtype == grad_out.dtype, lambda: f"expected dtype {self.dtype} for `grad_output` but got dtype {grad_out.dtype}", ) return self.new_empty(self.shape)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/__init__.py
colossalai/fx/__init__.py
from ._compatibility import compatibility, is_compatible_with_meta from .graph_module import ColoGraphModule from .passes import MetaInfoProp, metainfo_trace from .tracer import ColoTracer, meta_trace, symbolic_trace
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/graph_module.py
colossalai/fx/graph_module.py
import os import warnings from pathlib import Path from typing import Any, Dict, Optional, Union import torch import torch.nn as nn from torch.nn.modules.module import _addindent try: from torch.fx.graph import Graph, PythonCode, _PyTreeCodeGen from torch.fx.graph_module import GraphModule, _exec_with_source, _forward_from_src, _WrappedCall from colossalai.fx.codegen.activation_checkpoint_codegen import ActivationCheckpointCodeGen COLOGM = True except: from torch.fx.graph import Graph from torch.fx.graph_module import GraphModule COLOGM = False if COLOGM: class ColoGraphModule(GraphModule): def __init__( self, root: Union[torch.nn.Module, Dict[str, Any]], graph: Graph, class_name: str = "GraphModule", ckpt_codegen: bool = True, ): if ckpt_codegen: graph.set_codegen(ActivationCheckpointCodeGen()) super().__init__(root, graph, class_name) def bind(self, ckpt_def, globals): """Bind function needed for correctly execute gm forward We need to bind checkpoint functions and saved_tensor_hooks functions to gm so that we could correctly execute gm forward Args: ckpt_def (_type_): definition before the forward function globals (_type_): global variables """ ckpt_code = "\n".join(ckpt_def) globals_copy = globals.copy() _exec_with_source(ckpt_code, globals_copy) func_list = [func for func in globals_copy.keys() if "checkpoint" in func or "pack" in func] for func in func_list: tmp_func = globals_copy[func] setattr(self, func, tmp_func.__get__(self, self.__class__)) del globals_copy[func] def recompile(self) -> PythonCode: """ Recompile this GraphModule from its ``graph`` attribute. This should be called after editing the contained ``graph``, otherwise the generated code of this ``GraphModule`` will be out of date. """ if isinstance(self._graph._codegen, _PyTreeCodeGen): self._in_spec = self._graph._codegen.pytree_info.in_spec self._out_spec = self._graph._codegen.pytree_info.out_spec python_code = self._graph.python_code(root_module="self") self._code = python_code.src # To split ckpt functions code and forward code _code_list = self._code.split("\n") _fwd_def = [item for item in _code_list if "def forward" in item][0] _fwd_idx = _code_list.index(_fwd_def) ckpt_def = _code_list[:_fwd_idx] self._code = "\n".join(_code_list[_fwd_idx:]) self.bind(ckpt_def, python_code.globals) cls = type(self) cls.forward = _forward_from_src(self._code, python_code.globals) # Determine whether this class explicitly defines a __call__ implementation # to wrap. If it does, save it in order to have wrapped_call invoke it. # If it does not, wrapped_call can use a dynamic call to super() instead. # In most cases, super().__call__ should be torch.nn.Module.__call__. # We do not want to hold a reference to Module.__call__ here; doing so will # bypass patching of torch.nn.Module.__call__ done while symbolic tracing. cls_call = cls.__call__ if "__call__" in vars(cls) else None if "_wrapped_call" not in vars(cls): cls._wrapped_call = _WrappedCall(cls, cls_call) # type: ignore[attr-defined] def call_wrapped(self, *args, **kwargs): return self._wrapped_call(self, *args, **kwargs) cls.__call__ = call_wrapped # reset self._code to original src, otherwise to_folder will be wrong self._code = python_code.src return python_code def to_folder(self, folder: Union[str, os.PathLike], module_name: str = "FxModule"): """Dumps out module to ``folder`` with ``module_name`` so that it can be imported with ``from <folder> import <module_name>`` Args: folder (Union[str, os.PathLike]): The folder to write the code out to module_name (str): Top-level name to use for the ``Module`` while writing out the code """ folder = Path(folder) Path(folder).mkdir(exist_ok=True) torch.save(self.state_dict(), folder / "state_dict.pt") tab = " " * 4 # we add import colossalai here model_str = f""" import torch from torch.nn import * import colossalai class {module_name}(torch.nn.Module): def __init__(self): super().__init__() """ def _gen_model_repr(module_name: str, module: torch.nn.Module) -> Optional[str]: safe_reprs = [ nn.Linear, nn.Conv1d, nn.Conv2d, nn.Conv3d, nn.BatchNorm1d, nn.BatchNorm2d, nn.BatchNorm3d, ] if type(module) in safe_reprs: return f"{module.__repr__()}" else: return None blobified_modules = [] for module_name, module in self.named_children(): module_str = _gen_model_repr(module_name, module) if module_str is None: module_file = folder / f"{module_name}.pt" torch.save(module, module_file) blobified_modules.append(module_name) module_repr = module.__repr__().replace("\r", " ").replace("\n", " ") module_str = f"torch.load(r'{module_file}') # {module_repr}" model_str += f"{tab*2}self.{module_name} = {module_str}\n" for buffer_name, buffer in self._buffers.items(): if buffer is None: continue model_str += f"{tab*2}self.register_buffer('{buffer_name}', torch.empty({list(buffer.shape)}, dtype={buffer.dtype}))\n" for param_name, param in self._parameters.items(): if param is None: continue model_str += f"{tab*2}self.{param_name} = torch.nn.Parameter(torch.empty({list(param.shape)}, dtype={param.dtype}))\n" model_str += f"{tab*2}self.load_state_dict(torch.load(r'{folder}/state_dict.pt'))\n" model_str += f"{_addindent(self.code, 4)}\n" module_file = folder / "module.py" module_file.write_text(model_str) init_file = folder / "__init__.py" init_file.write_text("from .module import *") if len(blobified_modules) > 0: warnings.warn( "Was not able to save the following children modules as reprs -" f"saved as pickled files instead: {blobified_modules}" ) else: class ColoGraphModule(GraphModule): def __init__(self, root: Union[torch.nn.Module, Dict[str, Any]], graph: Graph, class_name: str = "GraphModule"): super().__init__(root, graph, class_name)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/proxy.py
colossalai/fx/proxy.py
from typing import Any import torch from torch.fx.proxy import Proxy from colossalai.fx.tracer.meta_patch import meta_patched_function __all__ = ["ColoProxy"] class ColoProxy(Proxy): """ ColoProxy is a proxy class which uses meta tensor to handle data-dependent control flow. The original torch.fx proxy cannot be used to infer the condition statement, with this proxy, torch.fx can still run even with if statements. Example:: proxy = tracer.create_proxy(...) proxy.meta_data = torch.empty(4, 2, device='meta') print(len(proxy)) # expect output 4 """ def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.node._meta_data = None @property def meta_data(self): return self.node._meta_data @meta_data.setter def meta_data(self, data: Any): self.node._meta_data = data @property def has_meta_data(self): return self._meta_data is not None def _assert_meta_data_is_tensor(self): assert ( torch.is_tensor(self._meta_data) and self._meta_data.is_meta ), f"Meta data is not a meta tensor for {self.node.name}" def _assert_has_meta_data(self): assert self._meta_data is not None, f"Meta data is not set for {self.node.name}" def __len__(self): self._assert_has_meta_data() return len(self.meta_data) def __int__(self): self._assert_has_meta_data() return int(self.meta_data) def __float__(self): self._assert_has_meta_data() return float(self.meta_data) def __bool__(self): self._assert_has_meta_data() return self.meta_data def __getattr__(self, k): return ColoAttribute(self, k) def __contains__(self, key): if self.node.op == "placeholder": # this is used to handle like # if x in kwargs # we don't handle this case for now return False return super().__contains__(key) def extract_meta(*args, **kwargs): """ This function is copied from _tracer_utils.py to avoid circular import issue. """ def _convert(val): if isinstance(val, ColoProxy): return val.meta_data elif isinstance(val, (list, tuple)): return type(val)([_convert(ele) for ele in val]) return val new_args = [_convert(val) for val in args] new_kwargs = {k: _convert(v) for k, v in kwargs.items()} return new_args, new_kwargs class ColoAttribute(ColoProxy): def __init__(self, root, attr: str): self.root = root self.attr = attr self.tracer = root.tracer self._node = None @property def node(self): if self._node is None: proxy = self.tracer.create_proxy("call_function", getattr, (self.root, self.attr), {}) if not isinstance(proxy, ColoProxy): meta_args, meta_kwargs = extract_meta(*(self.root, self.attr)) meta_out = getattr(*meta_args, **meta_kwargs) proxy = ColoProxy(proxy.node) proxy.meta_data = meta_out self._node = proxy.node return self._node def __call__(self, *args, **kwargs): proxy = self.tracer.create_proxy("call_method", self.attr, (self.root,) + args, kwargs) if not isinstance(proxy, ColoProxy): meta_args, meta_kwargs = extract_meta(*((self.root,) + args), **kwargs) method = getattr(meta_args[0].__class__, self.attr) if meta_patched_function.has(method): meta_target = meta_patched_function.get(method) elif meta_patched_function.has(method.__name__): meta_target = meta_patched_function.get(method.__name__) else: meta_target = method meta_out = meta_target(*meta_args, **meta_kwargs) proxy = ColoProxy(proxy.node) proxy.meta_data = meta_out return proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/memory_utils.py
colossalai/fx/profiler/memory_utils.py
from typing import Dict, List, Tuple, Union import torch from torch.fx import Node from .._compatibility import compatibility, is_compatible_with_meta __all__ = ["activation_size", "parameter_size", "is_inplace"] @compatibility(is_backward_compatible=True) def activation_size(out: Union[torch.Tensor, Dict, List, Tuple, int]) -> int: """Calculate activation size of a node. Args: activation (Union[torch.Tensor, Dict, List, Tuple, int]): The activation of a `torch.nn.Module` or `torch.nn.functional`. Returns: int: The activation size, unit is byte. """ act_size = 0 if isinstance(out, torch.Tensor): if out.is_quantized: act_size += out.numel() * torch._empty_affine_quantized([], dtype=out.dtype).element_size() else: act_size += out.numel() * torch.tensor([], dtype=out.dtype).element_size() elif isinstance(out, dict): value_list = [v for _, v in out.items()] act_size += activation_size(value_list) elif isinstance(out, tuple) or isinstance(out, list) or isinstance(out, set): for element in out: act_size += activation_size(element) return act_size @compatibility(is_backward_compatible=True) def parameter_size(mod: torch.nn.Module) -> int: """Calculate parameter size of a node. Args: mod (torch.nn.Module): The target `torch.nn.Module`. Returns: int: The parameter size, unit is byte. """ param_size = 0 for param in mod.parameters(): param_size += param.numel() * torch.tensor([], dtype=param.dtype).element_size() return param_size def is_inplace(n: Node): """Get the inplace argument from torch.fx.Node Args: node (Node): torch.fx.Node Returns: bool: indicates whether this op is inplace """ inplace = False if n.op == "call_function": inplace = n.kwargs.get("inplace", False) if is_compatible_with_meta(): from .constants import ALIAS_ATEN if n.target in ALIAS_ATEN: inplace = True elif n.op == "call_module": inplace = getattr(n.graph.owning_module.get_submodule(n.target), "inplace", False) return inplace
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/constants.py
colossalai/fx/profiler/constants.py
import torch __all__ = ["ALIAS_ATEN", "INPLACE_NEW", "INPLACE_MATH_ATEN", "CLONE_ATEN", "RELU_LIKE_OPS", "RELU_LIKE_MOD"] aten = torch.ops.aten ALIAS_ATEN = [ aten.detach.default, aten.t.default, aten.transpose.int, aten.view.default, aten._unsafe_view.default, aten._reshape_alias.default, ] INPLACE_NEW = [ aten.empty_like.default, aten.new_empty_strided.default, ] INPLACE_MATH_ATEN = [ aten.add_.Tensor, aten.sub_.Tensor, aten.div_.Tensor, aten.div_.Scalar, aten.mul_.Tensor, aten.bernoulli_.float, ] CLONE_ATEN = [ aten.clone.default, ] # See illustrations in # https://github.com/hpcaitech/ColossalAI/blob/main/colossalai/fx/profiler/constants.py OUTPUT_SAVED_OPS = [ torch.nn.functional.relu, torch.nn.functional.softmax, ] OUTPUT_SAVED_MOD = [ torch.nn.ReLU, torch.nn.Softmax, ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/profiler.py
colossalai/fx/profiler/profiler.py
import time from functools import partial from typing import Any, Callable, Dict, Tuple import torch from torch.fx import Graph, Node from torch.fx.node import Argument, Target from torch.nn.parameter import Parameter from torch.utils._pytree import tree_map from .._compatibility import compatibility from .constants import ALIAS_ATEN, OUTPUT_SAVED_MOD, OUTPUT_SAVED_OPS from .dataflow import GraphInfo, Phase, autograd_graph_analysis, is_phase from .memory_utils import activation_size, parameter_size from .opcount import flop_mapping from .tensor import MetaTensor __all__ = ["profile_function", "profile_module", "profile_method"] # super-dainiu: this cache should be global, otherwise it cannot # track duplicated tensors between nodes cache = set() # a global identifier for inplace ops do_not_cache = False def normalize_tuple(x): if not isinstance(x, tuple): return (x,) return x def is_autogradable(x): return isinstance(x, torch.Tensor) and x.is_floating_point() def detach_variables(x): if isinstance(x, torch.Tensor): requires_grad = x.requires_grad x = x.detach() x.requires_grad = requires_grad return x @compatibility(is_backward_compatible=True) def _profile_concrete(target: Callable, *args, **kwargs) -> Tuple[Tuple[Any, ...], GraphInfo]: """Profile a Callable function with args and kwargs on concrete devices by https://github.com/Cypher30 To profile the actual forward memory, we first run target in the context torch.no_grad() to get the fwd_mem_out, then we run target with grad enable to found the extra memory stored in the memory by memory allocated minus the fwd_mem_out. To profile the actual backward memory, we first make dummy gradient for torch.autograd.backward, then find the bwd_mem_tmp with memory peak during the process minus bwd_mem_out(it is actually equal to size of args and kwargs). We also add time stamps to profile the real forward and backward time. Args: target (Callable): A Callable function args (Any): Arguments kwargs (Any): Arguments Returns: Tuple[Tuple[Any, ...], GraphInfo]: Output for next node & memory cost and real forward and backward time. """ graphinfo = GraphInfo() # detach input from the graph args = tree_map(detach_variables, args) kwargs = tree_map(detach_variables, kwargs) if isinstance(target, str): # args[0] is the `self` object for this method call self_obj, *args_tail = args # calculate fwd_mem_out mem_stamp0 = torch.cuda.memory_allocated() with torch.no_grad(): out = getattr(self_obj, target)(*args_tail, **kwargs) mem_stamp1 = torch.cuda.memory_allocated() graphinfo.fwd_mem_out = mem_stamp1 - mem_stamp0 del out # calculate fwd_mem_tmp & fwd_time mem_stamp0 = torch.cuda.memory_allocated() fwd_time0 = time.time() out = getattr(self_obj, target)(*args_tail, **kwargs) fwd_time1 = time.time() graphinfo.fwd_time = fwd_time1 - fwd_time0 mem_stamp1 = torch.cuda.memory_allocated() graphinfo.fwd_mem_tmp = mem_stamp1 - mem_stamp0 - graphinfo.fwd_mem_out # calculate bwd_mem_tmp & bwd_time grad_tensors = tree_map(lambda x: torch.ones_like(x) if isinstance(x, torch.Tensor) else None, out) torch.cuda.reset_peak_memory_stats() mem_stamp0 = torch.cuda.memory_allocated() bwd_time0 = time.time() torch.autograd.backward(out, grad_tensors=grad_tensors) bwd_time1 = time.time() graphinfo.bwd_time = bwd_time1 - bwd_time0 mem_stamp1 = torch.cuda.max_memory_allocated() # calculate bwd memory stats # NOTE: the module should add param to bwd_mem_out for bwd_mem_tmp calculation graphinfo.bwd_mem_out = activation_size(args) + activation_size(kwargs) graphinfo.bwd_mem_out += parameter_size(target.__self__) if hasattr(target.__self__, "parameters") else 0 graphinfo.bwd_mem_tmp = mem_stamp1 - mem_stamp0 - graphinfo.bwd_mem_out else: # calculate fwd_mem_out mem_stamp0 = torch.cuda.memory_allocated() with torch.no_grad(): out = target(*args, **kwargs) mem_stamp1 = torch.cuda.memory_allocated() graphinfo.fwd_mem_out = mem_stamp1 - mem_stamp0 del out # calculate fwd_mem_tmp & fwd_time mem_stamp0 = torch.cuda.memory_allocated() fwd_time0 = time.time() out = target(*args, **kwargs) fwd_time1 = time.time() graphinfo.fwd_time = fwd_time1 - fwd_time0 mem_stamp1 = torch.cuda.memory_allocated() graphinfo.fwd_mem_tmp = mem_stamp1 - mem_stamp0 - graphinfo.fwd_mem_out # calculate bwd_mem_tmp & bwd_time grad_tensors = tree_map(lambda x: torch.ones_like(x) if isinstance(x, torch.Tensor) else None, out) torch.cuda.reset_peak_memory_stats() mem_stamp0 = torch.cuda.memory_allocated() bwd_time0 = time.time() torch.autograd.backward(out, grad_tensors=grad_tensors) bwd_time1 = time.time() graphinfo.bwd_time = bwd_time1 - bwd_time0 mem_stamp1 = torch.cuda.max_memory_allocated() # calculate bwd memory stats # NOTE: the module should add param to bwd_mem_out for bwd_mem_tmp calculation graphinfo.bwd_mem_out = activation_size(args) + activation_size(kwargs) graphinfo.bwd_mem_out += parameter_size(target.__self__) if hasattr(target.__self__, "parameters") else 0 graphinfo.bwd_mem_tmp = mem_stamp1 - mem_stamp0 - graphinfo.bwd_mem_out return tree_map(detach_variables, out), graphinfo @compatibility(is_backward_compatible=False) def _profile_meta(target: Callable, *args, **kwargs) -> Tuple[Tuple[Any, ...], GraphInfo]: """ Profile a Callable function with args and kwargs on meta devices. Args: target (Callable): A Callable function args (Any): Argument kwargs (Any): Argument Returns: out (Tuple[Any, ...]): The argument value that was retrieved. meta_info (GraphInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ # This subgraph traces aten level ops inside one node. subgraph = Graph() # `flop_count`` serves as a global dictionary to store results. flop_count = { Phase.FORWARD: 0, Phase.BACKWARD: 0, } # FlopTensor not only get the flop statistics of a single node, # it also build a full autograd graph for this node. # This makes sure we can analyze the dependencies of memory, and # decide which forward intermediate results should be kept until # backward is executed. # Hopefully, this attempt will provide a better estimation of memory. class FlopTensor(MetaTensor): _node: Node = None def __repr__(self): if self.grad_fn: return f"FlopTensor({self._tensor}, fake_device='{self.device}', size={tuple(self.shape)}, grad_fn={self.grad_fn})" return f"FlopTensor({self._tensor}, fake_device='{self.device}', size={tuple(self.shape)}, requires_grad={self.requires_grad})" @classmethod def __torch_dispatch__(cls, func, types, args=(), kwargs=None): args_node = tree_map(lambda x: x._node if isinstance(x, FlopTensor) else None, args) kwargs_node = tree_map(lambda x: x._node if isinstance(x, FlopTensor) else None, kwargs) node = subgraph.create_node("call_function", func, args_node, kwargs_node) out = super().__torch_dispatch__(func, types, args, kwargs) flop_count[phase] += flop_mapping[func](args, normalize_tuple(out)) node.meta["phase"] = phase # super-dainiu: in `nn.MultiheadAttention` this weird thing occurs, # i.e. `Phase.PLACEHOLDER` tensors are aliased and saved during # `Phase.FORWARD` if phase == Phase.FORWARD: if all(map(partial(is_phase, phase=Phase.PLACEHOLDER), node.all_input_nodes)) and func in ALIAS_ATEN: node.meta["phase"] = Phase.PLACEHOLDER # TODO(yby): specify `saved_tensors` for backward memory estimation node.meta["saved_tensor"] = [] if phase == Phase.BACKWARD: node.meta["saved_tensor"] = normalize_tuple(out) def wrap(x): if isinstance(x, MetaTensor): x = FlopTensor(x) x._node = node return x out = tree_map(wrap, out) return out def wrap(x): if isinstance(x, torch.Tensor): x = FlopTensor(x) if is_autogradable(x): x.requires_grad_(True) x._node = subgraph.create_node( "placeholder", "placeholder", (subgraph._root,), name=subgraph._graph_namespace.create_name("input", x._tensor), ) x._node.meta["phase"] = Phase.PLACEHOLDER x._node.meta["saved_tensor"] = [] return x # Basically, we need to detach the args and kwargs from the outer graph. args = tree_map(wrap, args) kwargs = tree_map(wrap, kwargs) def pack(x): global cache, do_not_cache if isinstance(x, FlopTensor) and not x._tensor.data_ptr() in cache: tensor = x._tensor.detach() tensor.data_ptr = x._tensor.data_ptr x._node.meta["saved_tensor"] += [tensor] if not do_not_cache: cache.add(x._tensor.data_ptr()) return x def unpack(x): return x # `phase` will mark the phase of autograd from outside scope. phase = Phase.FORWARD # mark saved tensors with saved_tensors_hooks with torch.autograd.graph.saved_tensors_hooks(pack, unpack): if isinstance(target, str): # args[0] is the `self` object for this method call self_obj, *args_tail = args out = getattr(self_obj, target)(*args_tail, **kwargs) else: out = target(*args, **kwargs) # If the output is not a floating point `torch.Tensor` or it does not # requires grad, then we should not run backward for this node. if all(map(lambda x: is_autogradable(x) and x.requires_grad, normalize_tuple(out))): grad_out = [torch.zeros_like(t) for t in normalize_tuple(out)] phase = Phase.BACKWARD torch.autograd.backward( out, grad_out, ) graph_info = autograd_graph_analysis(subgraph) graph_info.fwd_flop, graph_info.bwd_flop = flop_count[Phase.FORWARD], flop_count[Phase.BACKWARD] def extract_tensor(x: Any): if isinstance(x, MetaTensor): tensor = x._tensor.detach() tensor.data_ptr = x._tensor.data_ptr return tensor if not isinstance(x, torch.finfo): return x graph_info.fwd_out = list(map(extract_tensor, normalize_tuple(out))) def unwrap(x): return MetaTensor(x) if isinstance(x, torch.Tensor) else x return tree_map(unwrap, out), graph_info @compatibility(is_backward_compatible=True) def profile_function(target: "Target", device: str = "meta") -> Callable: """ Wrap a `call_function` node or `torch.nn.functional` in order to record the memory cost and FLOPs of the execution. Warnings: You may only use tensors with `device=meta` for this wrapped function. Only original `torch.nn.functional` are available. Examples: >>> input = torch.rand(100, 100, 100, 100, device='meta') >>> func = torch.nn.functional.relu >>> output, meta_info = profile_function(func)(input) """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: # find the grad for parameter in args and kwargs param_size = 0 def get_param_size(x): nonlocal param_size if isinstance(x, Parameter): param_size += activation_size(x) tree_map(get_param_size, args) tree_map(get_param_size, kwargs) # If there is an argument that this `call_function` is inplace, we should # still run the profiling but discard some results regarding `target` global do_not_cache inplace = kwargs.get("inplace", False) if target in OUTPUT_SAVED_OPS: do_not_cache = True if inplace: do_not_cache = True kwargs["inplace"] = False if device == "meta": out, meta = _profile_meta(func, *args, **kwargs) else: out, meta = _profile_concrete(func, *args, **kwargs) if inplace: kwargs["inplace"] = True meta.bwd_mem_tmp = 0 meta.bwd_mem_out = 0 do_not_cache = False meta.bwd_mem_out -= param_size return out, meta f.__name__ = target.__name__ func = target return f @compatibility(is_backward_compatible=True) def profile_method(target: "Target", device: str = "meta") -> Callable: """ Wrap a `call_method` node record the memory cost and FLOPs of the execution. """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: # execute the method and return the result assert isinstance(target, str), f"{target} instance is not str." if device == "meta": out, meta = _profile_meta(target, *args, **kwargs) else: out, meta = _profile_concrete(target, *args, **kwargs) return out, meta return f @compatibility(is_backward_compatible=True) def profile_module(module: torch.nn.Module, device: str = "meta") -> Callable: """ Wrap a `call_module` node or `torch.nn` in order to record the memory cost and FLOPs of the execution. Warnings: You may only use tensors with `device=meta` for this wrapped function. Only original `torch.nn` are available. Example: >>> input = torch.rand(4, 3, 224, 224, device='meta') >>> mod = torch.nn.Conv2d(3, 128, 3) >>> output, meta_info = profile_module(mod)(input) """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: # calculate parameter size param_size = parameter_size(module) # If there is an argument that this `call_module` is inplace, we should # still run the profiling but discard some results regarding `module`. global do_not_cache inplace = getattr(module, "inplace", False) if type(module) in OUTPUT_SAVED_MOD: do_not_cache = True if inplace: do_not_cache = True module.inplace = False if device == "meta": out, meta = _profile_meta(func, *args, **kwargs) else: out, meta = _profile_concrete(func, *args, **kwargs) if inplace: module.inplace = True meta.bwd_mem_tmp = 0 meta.bwd_mem_out = 0 do_not_cache = False # grad for param will not be counted meta.bwd_mem_out -= param_size return out, meta f.__name__ = module.__class__.__name__ func = module.forward return f
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/opcount.py
colossalai/fx/profiler/opcount.py
# adopted from https://github.com/facebookresearch/fvcore/blob/main/fvcore/nn/jit_handles.py # ideas from https://pastebin.com/AkvAyJBw import operator from functools import partial, reduce from numbers import Number from typing import Any, Callable, List import torch from packaging import version aten = torch.ops.aten def matmul_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for matmul. """ # Inputs should be a list of length 2. # Inputs contains the shapes of two matrices. input_shapes = [v.shape for v in inputs] assert len(input_shapes) == 2, input_shapes # There are three cases: 1) gemm, 2) gemv, 3) dot if all(len(shape) == 2 for shape in input_shapes): # gemm assert input_shapes[0][-1] == input_shapes[1][-2], input_shapes elif all(len(shape) == 1 for shape in input_shapes): # dot assert input_shapes[0][0] == input_shapes[1][0], input_shapes # expand shape input_shapes[0] = torch.Size([1, input_shapes[0][0]]) input_shapes[1] = torch.Size([input_shapes[1][0], 1]) else: # gemv if len(input_shapes[0]) == 1: assert input_shapes[0][0] == input_shapes[1][-2], input_shapes input_shapes.reverse() else: assert input_shapes[1][0] == input_shapes[0][-1], input_shapes # expand the shape of the vector to [batch size, 1] input_shapes[-1] = torch.Size([input_shapes[-1][-1], 1]) flops = reduce(operator.mul, input_shapes[0]) * input_shapes[-1][-1] return flops def addmm_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for fully connected layers. """ # Count flop for nn.Linear # inputs is a list of length 3. input_shapes = [v.shape for v in inputs[1:3]] # input_shapes[0]: [batch size, input feature dimension] # input_shapes[1]: [input feature dimension, output feature dimension] assert len(input_shapes[0]) == 2, input_shapes[0] assert len(input_shapes[1]) == 2, input_shapes[1] batch_size, input_dim = input_shapes[0] output_dim = input_shapes[1][1] flops = batch_size * input_dim * output_dim return flops def linear_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for the aten::linear operator. """ # Inputs is a list of length 3; unlike aten::addmm, it is the first # two elements that are relevant. input_shapes = [v.shape for v in inputs[0:2]] # input_shapes[0]: [dim0, dim1, ..., input_feature_dim] # input_shapes[1]: [output_feature_dim, input_feature_dim] assert input_shapes[0][-1] == input_shapes[1][-1] flops = reduce(operator.mul, input_shapes[0]) * input_shapes[1][0] return flops def bmm_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for the bmm operation. """ # Inputs should be a list of length 2. # Inputs contains the shapes of two tensor. assert len(inputs) == 2, len(inputs) input_shapes = [v.shape for v in inputs] n, c, t = input_shapes[0] d = input_shapes[-1][-1] flops = n * c * t * d return flops def baddbmm_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for the baddbmm(batch add and batch matmul) operation. """ # Inputs = [input, batch1, batch2] # out = input + batch1 x batch2 assert len(inputs) == 3, len(inputs) n, c, t = inputs[1].shape d = inputs[2].shape[-1] flops = n * c * t * d return flops def conv_flop_count( x_shape: List[int], w_shape: List[int], out_shape: List[int], transposed: bool = False, ) -> Number: """ Count flops for convolution. Note only multiplication is counted. Computation for addition and bias is ignored. Flops for a transposed convolution are calculated as flops = (x_shape[2:] * prod(w_shape) * batch_size). Args: x_shape (list(int)): The input shape before convolution. w_shape (list(int)): The filter shape. out_shape (list(int)): The output shape after convolution. transposed (bool): is the convolution transposed Returns: int: the number of flops """ batch_size = x_shape[0] conv_shape = (x_shape if transposed else out_shape)[2:] flops = batch_size * reduce(operator.mul, w_shape) * reduce(operator.mul, conv_shape) return flops def conv_flop_jit(inputs: List[Any], outputs: List[Any]): """ Count flops for convolution. """ x, w = inputs[:2] x_shape, w_shape, out_shape = (x.shape, w.shape, outputs[0].shape) transposed = inputs[6] return conv_flop_count(x_shape, w_shape, out_shape, transposed=transposed) def transpose_shape(shape): return [shape[1], shape[0]] + list(shape[2:]) def conv_backward_flop_jit(inputs: List[Any], outputs: List[Any]): grad_out_shape, x_shape, w_shape = [i.shape for i in inputs[:3]] output_mask = inputs[-1] fwd_transposed = inputs[7] flop_count = 0 if output_mask[0]: grad_input_shape = outputs[0].shape flop_count += conv_flop_count(grad_out_shape, w_shape, grad_input_shape, not fwd_transposed) if output_mask[1]: grad_weight_shape = outputs[1].shape flop_count += conv_flop_count(transpose_shape(x_shape), grad_out_shape, grad_weight_shape, fwd_transposed) return flop_count def norm_flop_counter(affine_arg_index: int, input_arg_index: int) -> Callable: """ Args: affine_arg_index: index of the affine argument in inputs """ def norm_flop_jit(inputs: List[Any], outputs: List[Any]) -> Number: """ Count flops for norm layers. """ # Inputs[0] contains the shape of the input. input_shape = inputs[input_arg_index].shape has_affine = ( inputs[affine_arg_index].shape is not None if hasattr(inputs[affine_arg_index], "shape") else inputs[affine_arg_index] ) assert 2 <= len(input_shape) <= 5, input_shape # 5 is just a rough estimate flop = reduce(operator.mul, input_shape) * (5 if has_affine else 4) return flop return norm_flop_jit def batchnorm_flop_jit(inputs: List[Any], outputs: List[Any], training: bool = None) -> Number: if training is None: training = inputs[-3] assert isinstance(training, bool), "Signature of aten::batch_norm has changed!" if training: return norm_flop_counter(1, 0)(inputs, outputs) # pyre-ignore has_affine = inputs[1].shape is not None input_shape = reduce(operator.mul, inputs[0].shape) return input_shape * (2 if has_affine else 1) def elementwise_flop_counter(input_scale: float = 1, output_scale: float = 0) -> Callable: """ Count flops by input_tensor.numel() * input_scale + output_tensor.numel() * output_scale Args: input_scale: scale of the input tensor (first argument) output_scale: scale of the output tensor (first element in outputs) """ def elementwise_flop(inputs: List[Any], outputs: List[Any]) -> Number: ret = 0 if input_scale != 0: shape = inputs[0].shape ret += input_scale * reduce(operator.mul, shape) if shape else 0 if output_scale != 0: shape = outputs[0].shape ret += output_scale * reduce(operator.mul, shape) if shape else 0 return ret return elementwise_flop def zero_flop_jit(*args): """ Count flops for zero flop layers. """ return 0 if version.parse(torch.__version__) >= version.parse("1.12.0") and version.parse(torch.__version__) < version.parse( "2.0.0" ): flop_mapping = { # gemm, gemv and dot aten.mm.default: matmul_flop_jit, aten.mv.default: matmul_flop_jit, aten.dot.default: matmul_flop_jit, aten.matmul.default: matmul_flop_jit, aten.addmm.default: addmm_flop_jit, aten.bmm.default: bmm_flop_jit, aten.baddbmm.default: baddbmm_flop_jit, # convolution aten.convolution.default: conv_flop_jit, aten._convolution.default: conv_flop_jit, aten.convolution_backward.default: conv_backward_flop_jit, # normalization aten.native_batch_norm.default: batchnorm_flop_jit, aten.native_batch_norm_backward.default: batchnorm_flop_jit, aten.cudnn_batch_norm.default: batchnorm_flop_jit, aten.cudnn_batch_norm_backward.default: partial(batchnorm_flop_jit, training=True), aten.native_layer_norm.default: norm_flop_counter(2, 0), aten.native_layer_norm_backward.default: norm_flop_counter(2, 0), aten.native_group_norm.default: norm_flop_counter(2, 0), aten.native_group_norm_backward.default: norm_flop_counter(2, 0), # pooling aten.avg_pool1d.default: elementwise_flop_counter(1, 0), aten.avg_pool2d.default: elementwise_flop_counter(1, 0), aten.avg_pool2d_backward.default: elementwise_flop_counter(0, 1), aten.avg_pool3d.default: elementwise_flop_counter(1, 0), aten.avg_pool3d_backward.default: elementwise_flop_counter(0, 1), aten.max_pool1d.default: elementwise_flop_counter(1, 0), aten.max_pool2d.default: elementwise_flop_counter(1, 0), aten.max_pool3d.default: elementwise_flop_counter(1, 0), aten.max_pool1d_with_indices.default: elementwise_flop_counter(1, 0), aten.max_pool2d_with_indices.default: elementwise_flop_counter(1, 0), aten.max_pool2d_with_indices_backward.default: elementwise_flop_counter(0, 1), aten.max_pool3d_with_indices.default: elementwise_flop_counter(1, 0), aten.max_pool3d_with_indices_backward.default: elementwise_flop_counter(0, 1), aten._adaptive_avg_pool2d.default: elementwise_flop_counter(1, 0), aten._adaptive_avg_pool2d_backward.default: elementwise_flop_counter(0, 1), aten._adaptive_avg_pool3d.default: elementwise_flop_counter(1, 0), aten._adaptive_avg_pool3d_backward.default: elementwise_flop_counter(0, 1), aten.embedding_dense_backward.default: elementwise_flop_counter(0, 1), aten.embedding.default: elementwise_flop_counter(1, 0), aten.upsample_nearest2d.vec: elementwise_flop_counter(0, 1), aten.upsample_nearest2d_backward.vec: elementwise_flop_counter(0, 1), } elementwise_flop_aten = [ # basic op aten.add.Tensor, aten.add_.Tensor, aten.div.Tensor, aten.div_.Tensor, aten.div.Scalar, aten.div_.Scalar, aten.mul.Tensor, aten.mul.Scalar, aten.mul_.Tensor, aten.neg.default, aten.pow.Tensor_Scalar, aten.rsub.Scalar, aten.sum.default, aten.sum.dim_IntList, aten.mean.dim, aten.sub.Tensor, aten.sub_.Tensor, aten.exp.default, aten.sin.default, aten.cos.default, # activation op aten.hardswish.default, aten.hardswish_.default, aten.hardswish_backward.default, aten.hardtanh.default, aten.hardtanh_.default, aten.hardtanh_backward.default, aten.hardsigmoid_backward.default, aten.hardsigmoid.default, aten.gelu.default, aten.gelu_backward.default, aten.silu.default, aten.silu_.default, aten.silu_backward.default, aten.sigmoid.default, aten.sigmoid_backward.default, aten._softmax.default, aten._softmax_backward_data.default, aten.relu_.default, aten.relu.default, aten.tanh.default, aten.tanh_backward.default, aten.threshold_backward.default, # dropout aten.native_dropout.default, aten.native_dropout_backward.default, ] for op in elementwise_flop_aten: flop_mapping[op] = elementwise_flop_counter(1, 0) # TODO: this will be removed in future zero_flop_aten = [ aten.as_strided.default, aten.as_strided_.default, aten.bernoulli_.float, aten.cat.default, aten.clone.default, aten.copy_.default, aten.detach.default, aten.expand.default, aten.empty_like.default, aten.new_empty.default, aten.new_empty_strided.default, aten.ones_like.default, aten._reshape_alias.default, aten.select.int, aten.select_backward.default, aten.squeeze.dim, aten.slice.Tensor, aten.slice_backward.default, aten.stack.default, aten.split.Tensor, aten.permute.default, aten.t.default, aten.transpose.int, aten._to_copy.default, aten.unsqueeze.default, aten.unbind.int, aten._unsafe_view.default, aten.view.default, aten.where.self, aten.zero_.default, aten.zeros_like.default, aten.fill_.Scalar, aten.stack.default, ] # yapf: disable for op in zero_flop_aten: flop_mapping[op] = zero_flop_jit else: flop_mapping = {} elementwise_flop_aten = {} zero_flop_aten = {}
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/dataflow.py
colossalai/fx/profiler/dataflow.py
from dataclasses import dataclass, field from enum import Enum from typing import Dict, List from torch.fx import Graph, Node from .._compatibility import compatibility from .memory_utils import activation_size, is_inplace class Phase(Enum): FORWARD = 0 BACKWARD = 1 PLACEHOLDER = 2 @compatibility(is_backward_compatible=True) @dataclass class GraphInfo: """ GraphInfo is a dataclass for MetaInfo, which measures the execution memory cost and FLOPs with `MetaTensor`. The dataflow analysis is conducted on a single node of the FX graph. ============================================================================ ------------------------------- | Node | [fwd_in] are ---> | [fwd_in] [bwd_out] | <----- [bwd_out] is marks the memory for `grad_out`. placeholders saved for | | \__________ | | backward. | | \ | | | [fwd_tmp] ------> [bwd_tmp] | <----- | | \_________ | | [bwd_tmp] marks the peak memory | / \ \ | | in backward pass. [x] is not counted ---> | [x] [fwd_tmp] -> [bwd_tmp] | <----- in [fwd_tmp] because | | \_____ | | it is not saved for | | \ | | backward. | [fwd_out] \ | | <----- [fwd_out] is [fwd_in] for the next node. ------------------------------- ============================================================================ Attributes: fwd_flop (int): The forward FLOPs of a certain node. fwd_time (float): The real forward time (s) of a certain node. bwd_flop (int): The backward FLOPs of a certain node. bwd_time (float): The real backward time (s) of a certain node. save_fwd_in (bool): The decision variable of whether to save the fwd_mem_out of parent nodes. fwd_in (List): See the above illustration. fwd_tmp (List): See the above illustration. fwd_out (List): See the above illustration. fwd_mem_tmp (int): See the above illustration. fwd_mem_out (int): See the above illustration. bwd_mem_tmp (int): See the above illustration. bwd_mem_out (int): See the above illustration. """ # TODO(super-dainiu): removed redundant items, currently all of them are necessary for development fwd_flop: int = 0 fwd_time: float = 0.0 bwd_flop: int = 0 bwd_time: float = 0.0 save_fwd_in: bool = False fwd_in: List = field(default_factory=list) fwd_tmp: List = field(default_factory=list) fwd_out: List = field(default_factory=list) fwd_mem_tmp: int = 0 fwd_mem_out: int = 0 bwd_mem_tmp: int = 0 bwd_mem_out: int = 0 def is_phase(n: Node, phase: Phase) -> bool: assert "phase" in n.meta, f"Node meta of {n} has no key `phase`!" return n.meta["phase"] == phase @compatibility(is_backward_compatible=False) def autograd_graph_analysis(graph: Graph) -> GraphInfo: """Analyze the autograd node dependencies and find out the memory usage. Basically the input graph should have all nodes marked for keyword `phase`. Nodes should have attribute `out` indicating the output of each node. ============================================================================ Placeholder ----> p o <---- We need to keep track of grad out |\________ | ↓ ↘| f --------> b |\ \_____ ↑ | \ ↘ / f f ----> b <---- Not every forward result needs to be saved for backward | \____ ↑ ↘ ↘| f ----> b <---- Backward can be freed as soon as it is required no more. ↘ ↗ l ============================================================================= Args: graph (Graph): The autograd graph with nodes marked for keyword `phase`. Returns: graph_info (GraphInfo): Meta information for the dataflow. """ def _peak_memory(deps: Dict[Node, int]): peak_mem = 0 for k, v in deps.items(): if v > 0 and is_phase(k, Phase.BACKWARD) and not all(map(is_inplace, k.users)) and not is_inplace(k): peak_mem += activation_size(k.meta["saved_tensor"]) if v <= float("-inf") and is_phase(k, Phase.FORWARD): peak_mem -= activation_size(k.meta["saved_tensor"]) return peak_mem # deps is used to track all the memory dependencies of the graph. deps = {} graph_info = GraphInfo() for n in graph.nodes: n: Node deps[n] = len(n.users) # A forward tensor who is marked `save` but is also # an input to `Phase.FORWARD` should be saved during forward. # If the tensor is a placeholder, then it belongs to `fwd_mem_in`. # Any `fwd_mem_in` should be kept in memory even this function # is checkpointed. # Otherwise, the tensor belongs to `fwd_mem_tmp`. If we checkpoint # the node, `fwd_mem_tmp` can be freed. if is_phase(n, Phase.PLACEHOLDER): graph_info.fwd_in += n.meta["saved_tensor"] if is_phase(n, Phase.FORWARD): graph_info.fwd_tmp += n.meta["saved_tensor"] elif is_phase(n, Phase.BACKWARD): if len(n.users): graph_info.bwd_mem_tmp = max(graph_info.bwd_mem_tmp, _peak_memory(deps)) else: # TODO: some of the bwd_mem_out might be model parameters. # basically a backward node without user is a `grad_out` node graph_info.bwd_mem_out += activation_size(n.meta["saved_tensor"]) for input_n in n.all_input_nodes: if input_n in deps: deps[input_n] -= 1 if deps[input_n] <= 0: deps[input_n] = float("-inf") return graph_info
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/__init__.py
colossalai/fx/profiler/__init__.py
from .._compatibility import is_compatible_with_meta if is_compatible_with_meta(): from .opcount import flop_mapping from .profiler import profile_function, profile_method, profile_module from .shard_utils import ( calculate_bwd_time, calculate_fwd_in, calculate_fwd_out, calculate_fwd_time, calculate_fwd_tmp, ) from .tensor import MetaTensor else: from .experimental import ( meta_profiler_function, meta_profiler_module, profile_function, profile_method, profile_module, calculate_fwd_in, calculate_fwd_tmp, calculate_fwd_out, ) from .dataflow import GraphInfo from .memory_utils import activation_size, is_inplace, parameter_size
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/tensor.py
colossalai/fx/profiler/tensor.py
import uuid import torch from torch.types import _device from torch.utils._pytree import tree_map from .._compatibility import compatibility from .constants import ALIAS_ATEN __all__ = ["MetaTensor"] def set_data_ptr(x): if isinstance(x, torch.Tensor): if not x.data_ptr(): data_ptr = uuid.uuid4() x.data_ptr = lambda: data_ptr @compatibility(is_backward_compatible=False) class MetaTensor(torch.Tensor): """ A wrapping tensor that hacks `torch.autograd` without patching more `torch.ops.aten` ops. `fake_device` is the device that `MetaTensor` is supposed to run on. """ _tensor: torch.Tensor @staticmethod def __new__(cls, elem, fake_device=None): # Avoid multiple wrapping if isinstance(elem, MetaTensor): fake_device = elem.device if fake_device is None else fake_device elem = elem._tensor # The wrapping tensor (MetaTensor) shouldn't hold any # memory for the class in question, but it should still # advertise the same device as before r = torch.Tensor._make_wrapper_subclass( cls, elem.size(), strides=elem.stride(), storage_offset=elem.storage_offset(), dtype=elem.dtype, layout=elem.layout, device=fake_device or (elem.device if elem.device.type != "meta" else torch.device("cpu")), requires_grad=elem.requires_grad, ) # deceive the frontend for aten selections r._tensor = elem # ...the real tensor is held as an element on the tensor. if not r._tensor.is_meta: r._tensor = r._tensor.to(torch.device("meta")) # only tensor not on `meta` should be copied to `meta` set_data_ptr(r._tensor) return r def __repr__(self): if self.grad_fn: return f"MetaTensor(..., size={tuple(self.shape)}, device='{self.device}', dtype={self.dtype}, grad_fn={self.grad_fn})" return f"MetaTensor(..., size={tuple(self.shape)}, device='{self.device}', dtype={self.dtype})" @classmethod def __torch_dispatch__(cls, func, types, args=(), kwargs=None): fake_device = None def unwrap(x): nonlocal fake_device if isinstance(x, MetaTensor): fake_device = x.device x = x._tensor elif isinstance(x, torch.Tensor): fake_device = x.device x = x.to(torch.device("meta")) return x args = tree_map(unwrap, args) kwargs = tree_map(unwrap, kwargs) if "device" in kwargs: fake_device = kwargs["device"] kwargs["device"] = torch.device("meta") # run aten for backend=CPU but actually on backend=Meta out = func(*args, **kwargs) # here we keep the uuid of input because ALIAS_ATEN do not generate a physical copy # of the input if func in ALIAS_ATEN: out.data_ptr = args[0].data_ptr # Now, we want to continue propagating this tensor, so we rewrap Tensors in # our custom tensor subclass def wrap(x): if isinstance(x, torch.Tensor): nonlocal fake_device if not x.is_meta: x = x.to(torch.device("meta")) return MetaTensor(x, fake_device=fake_device) if isinstance(x, torch.Tensor) else x return tree_map(wrap, out) def to(self, *args, **kwargs) -> torch.Tensor: """An extension of `torch.Tensor.to()` to MetaTensor Returns: result (MetaTensor): MetaTensor Usage: >>> tensor = MetaTensor(torch.rand(10), fake_device='cuda:100') >>> tensor.to(torch.uint8) MetaTensor(tensor(..., device='meta', size=(10,), dtype=torch.uint8), fake_device='cuda:100') >>> tensor.to(torch.device('cuda:42')) MetaTensor(tensor(..., device='meta', size=(10,)), fake_device='cuda:42') >>> tensor.to('vulkan') MetaTensor(tensor(..., device='meta', size=(10,)), fake_device='vulkan') """ # this imitates c++ function in the way of @overload fake_device = None def replace(x): nonlocal fake_device if isinstance(x, str) or isinstance(x, _device): fake_device = x return "meta" return x elem = self._tensor.to(*tree_map(replace, args), **tree_map(replace, kwargs)) return MetaTensor(elem, fake_device=fake_device) def cpu(self, *args, **kwargs): if self.device.type == "cpu": return self.to(*args, **kwargs) return self.to(*args, device="cpu", **kwargs) def cuda(self, device=None, non_blocking=False): if device is not None: return self.to(device=device, non_blocking=non_blocking) return self.to(device="cuda:0", non_blocking=non_blocking)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/shard_utils.py
colossalai/fx/profiler/shard_utils.py
import torch from torch.fx import Node from .._compatibility import compatibility, is_compatible_with_meta from .memory_utils import activation_size if is_compatible_with_meta(): from .constants import OUTPUT_SAVED_MOD, OUTPUT_SAVED_OPS __all__ = ["calculate_fwd_in", "calculate_fwd_tmp", "calculate_fwd_out"] @compatibility(is_backward_compatible=False) def calculate_fwd_in(n: Node) -> int: """A helper function to calculate `fwd_in` (with sharding spec) Args: n (Node): a node from the graph Returns: fwd_in (int): the result of `fwd_in` """ # TODO(super-dainiu): should divide the memory by sharding spec return activation_size(n.meta["fwd_in"]) @compatibility(is_backward_compatible=False) def calculate_fwd_tmp(n: Node) -> int: """A helper function to calculate `fwd_tmp` (with sharding spec) Currently, `torch.nn.ReLU` behaves weirdly, so we have to patch it for accuracy. Args: n (Node): a node from the graph Returns: fwd_tmp (int): the result of `fwd_tmp` """ # TODO(super-dainiu): should divide the memory by sharding spec def is_relu_like_node(n: Node) -> bool: """Check if a node is a ReLU-like node. ReLU-like nodes have the following properties: - They are either `call_function` or `call_module` - Their output tensors are directly saved for backward - Their input tensors are not saved for backward An example is `torch.nn.functional.softmax` which has (forward + backward): def forward(self, input_2): _softmax_default = torch.ops.aten._softmax.default(input_2, None, None); input_2 = None zeros_like_default = torch.ops.aten.zeros_like.default(_softmax_default, dtype = None, layout = None, device = None, pin_memory = None) detach_default = torch.ops.aten.detach.default(_softmax_default); _softmax_default = None _softmax_backward_data_default = torch.ops.aten._softmax_backward_data.default(zeros_like_default, detach_default, None, None); zeros_like_default = detach_default = None detach_default_1 = torch.ops.aten.detach.default(_softmax_backward_data_default); _softmax_backward_data_default = None detach_default_2 = torch.ops.aten.detach.default(detach_default_1); detach_default_1 = None Args: n (Node): A node from the graph Returns: bool: Whether the node is a ReLU-like node """ if n.op == "call_function": return n.target in OUTPUT_SAVED_OPS elif n.op == "call_module": return type(n.graph.owning_module.get_submodule(n.target)) in OUTPUT_SAVED_MOD return False if not is_relu_like_node(n): return activation_size(n.meta["fwd_tmp"]) return 0 @compatibility(is_backward_compatible=False) def calculate_fwd_out(n: Node) -> int: """A helper function to calculate `fwd_out` (with sharding spec) Args: n (Node): a node from the graph Returns: fwd_out (int): the result of `fwd_out` """ # TODO(super-dainiu): should divide the memory by sharding spec def intersect(a, b): return {k: a[k] for k in a if k in b} fwd_in = dict() for u in n.users: fwd_in.update({x.data_ptr(): x for x in u.meta["fwd_in"] if isinstance(x, torch.Tensor)}) fwd_out = {x.data_ptr(): x for x in n.meta["fwd_out"] if isinstance(x, torch.Tensor)} return activation_size(intersect(fwd_in, fwd_out)) def calculate_fwd_time(n: Node) -> float: """A helper function to calculate `fwd_time` (with sharding spec) Args: n (Node): a node from the graph Returns: fwd_time (float): the result of `fwd_time` """ # TODO(super-dainiu): should divide the time by the number of GPUs as well as TFLOPs return n.meta["fwd_time"] def calculate_bwd_time(n: Node) -> float: """A helper function to calculate `bwd_time` (with sharding spec) Args: n (Node): a node from the graph Returns: bwd_time (float): the result of `bwd_time` """ # TODO(super-dainiu): should divide the time by the number of GPUs as well as TFLOPs return n.meta["bwd_time"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/registry.py
colossalai/fx/profiler/experimental/registry.py
class ProfilerRegistry: def __init__(self, name): self.name = name self.store = {} def register(self, source): def wrapper(func): self.store[source] = func return func return wrapper def get(self, source): assert source in self.store target = self.store[source] return target def has(self, source): return source in self.store meta_profiler_function = ProfilerRegistry(name="patched_functions_for_meta_profile") meta_profiler_module = ProfilerRegistry(name="patched_modules_for_meta_profile")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/constants.py
colossalai/fx/profiler/experimental/constants.py
from operator import add, floordiv, getitem, mul, neg, pos, setitem, sub import torch __all__ = ["INPLACE_OPS", "INPLACE_METHOD", "NON_INPLACE_METHOD"] # TODO fill out the inplace ops INPLACE_OPS = [ add, sub, mul, floordiv, neg, pos, getitem, setitem, getattr, torch.Tensor.cpu, ] # TODO: list all call_methods that are inplace here INPLACE_METHOD = [ "transpose", "permute", # TODO: reshape may return a copy of the data if the data is not contiguous "reshape", "dim", "flatten", "size", "view", "unsqueeze", "to", "type", "flatten", ] # TODO: list all call_methods that are not inplace here NON_INPLACE_METHOD = [ "chunk", "contiguous", "expand", "mean", "split", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler.py
colossalai/fx/profiler/experimental/profiler.py
from dataclasses import dataclass from typing import Any, Callable, Dict, Tuple import torch from torch.fx.node import Argument, Target from ..._compatibility import compatibility from ..memory_utils import activation_size from .constants import INPLACE_METHOD, INPLACE_OPS, NON_INPLACE_METHOD from .registry import meta_profiler_function, meta_profiler_module __all__ = ["profile_function", "profile_module", "profile_method"] # this is for compatibility use @compatibility(is_backward_compatible=True) @dataclass class GraphInfo: """ GraphInfo is a dataclass for MetaInfo, which measures the execution memory cost and FLOPs with `MetaTensor`. The dataflow analysis is conducted on a single node of the FX graph. ============================================================================ ------------------------------- | Node | [fwd_in] are ---> | [fwd_in] [bwd_out] | <----- [bwd_out] is marks the memory for `grad_out` placeholders saved for | | \__________ | | backward. | | \ | | | [fwd_tmp] ------> [bwd_tmp] | <----- | | \_________ | | [bwd_tmp] marks the peak memory | / \ \ | | in backward pass. [x] is not counted ---> | [x] [fwd_tmp] -> [bwd_tmp] | <----- in [fwd_tmp] because | | | \_____ | | it is not saved for | | | \ | | backward. ------------------------------- ============================================================================ Attributes: fwd_flop (int): The forward FLOPs of a certain node bwd_flop (int): The backward FLOPs of a certain node. fwd_mem_in (int): See the above illustration. fwd_mem_tmp (int): See the above illustration. bwd_mem_tmp (int): See the above illustration. bwd_mem_out (int): See the above illustration. """ fwd_flop: int = 0 bwd_flop: int = 0 fwd_mem_in: int = 0 fwd_mem_tmp: int = 0 bwd_mem_tmp: int = 0 bwd_mem_out: int = 0 CALL_FUNCTION_MSG = """ Colossal-AI hasn't supported profiling for {}, you might manually patch it with the following code.\n from colossalai.fx.profiler.experimental import meta_profiler_function @meta_profiler_function.register(YOUR_FUNCTION) def profile_YOUR_FUNCTION(input: torch.Tensor, *args) -> Tuple[int, int]: flops = ... macs = ... return flops, macs """ CALL_METHOD_MSG = "Please check if {} is an inplace method. If so, add target to INPLACE_METHOD={}. Otherwise, add target to NON_INPLACE_METHOD={}" CALL_MODULE_MSG = """ Colossal-AI hasn't supported profiling for {}, you might manually patch it with the following code.\n from colossalai.fx.profiler.experimental import meta_profiler_module @meta_profiler_module.register(YOUR_MODULE) def profile_YOUR_MODULE(self: torch.nn.Module, input: torch.Tensor) -> Tuple[int, int]: flops = ... macs = ... return flops, macs """ @compatibility(is_backward_compatible=True) def profile_function(target: "Target") -> Callable: """ Wrap a `call_function` node or `torch.nn.functional` in order to record the memory cost and FLOPs of the execution. Unfortunately, backward memory cost and FLOPs are estimated results. Warnings: You may only use tensors with `device=meta` for this wrapped function. Only original `torch.nn.functional` are available. Examples: >>> input = torch.rand(100, 100, 100, 100, device='meta') >>> func = torch.nn.functional.relu >>> output, (fwd_flop, bwd_flop), (fwd_tmp, fwd_out, bwd_tmp, bwd_out) = profile_function(func)(input, inplace=False) """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: assert meta_profiler_function.has(target) or meta_profiler_function.has( target.__name__ ), CALL_FUNCTION_MSG.format(target) fwd_tmp = 0 fwd_out = 0 out = func(*args, **kwargs) if target not in INPLACE_OPS and not kwargs.get("inplace", False): fwd_out = activation_size(out) if meta_profiler_function.has(target): profiler = meta_profiler_function.get(target) else: profiler = meta_profiler_function.get(target.__name__) fwd_flop, _ = profiler(*args, **kwargs) return out, GraphInfo(fwd_flop, fwd_flop * 2, fwd_tmp, fwd_out, fwd_tmp + fwd_out, 0) f.__name__ = target.__name__ func = target return f @compatibility(is_backward_compatible=True) def profile_method(target: "Target") -> Callable: """ Wrap a `call_method` node record the memory cost and FLOPs of the execution. Warnings: This is not fully implemented and you may follow the error message to debug. """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: # args[0] is the `self` object for this method call self_obj, *args_tail = args # execute the method and return the result assert isinstance(target, str), f"{target} instance is not str." out = getattr(self_obj, target)(*args_tail, **kwargs) assert target in INPLACE_METHOD + NON_INPLACE_METHOD, CALL_METHOD_MSG.format( target, INPLACE_METHOD, NON_INPLACE_METHOD ) # call_method has no parameters and are MOSTLY(?) inplace, and has no FLOPs or MACs. fwd_tmp = 0 if target in INPLACE_METHOD else activation_size(out) fwd_out = 0 if target not in INPLACE_METHOD else activation_size(out) return out, GraphInfo(0, 0, fwd_tmp, fwd_out, fwd_tmp + fwd_out, 0) return f @compatibility(is_backward_compatible=True) def profile_module(module: torch.nn.Module) -> Callable: """ Wrap a `call_module` node or `torch.nn` in order to record the memory cost and FLOPs of the execution. Warnings: You may only use tensors with `device=meta` for this wrapped function. Only original `torch.nn` are available. Example: >>> input = torch.rand(4, 3, 224, 224, device='meta') >>> mod = torch.nn.Conv2d(3, 128, 3) >>> output, (fwd_flop, bwd_flop), (fwd_tmp, fwd_out, bwd_tmp, bwd_out) = profile_module(mod)(input) """ def f(*args: Tuple[Argument, ...], **kwargs: Dict[str, Any]) -> Any: assert meta_profiler_module.has(type(module)), CALL_MODULE_MSG.format(type(module)) fwd_tmp = 0 fwd_out = 0 out = func(*args, **kwargs) if getattr(module, "inplace", False): fwd_out = activation_size(out) profiler = meta_profiler_module.get(type(module)) fwd_flop, _ = profiler(module, *args, **kwargs) return out, GraphInfo(fwd_flop, fwd_flop * 2, fwd_tmp, fwd_out, fwd_tmp + fwd_out, 0) f.__name__ = module.__class__.__name__ func = module.forward return f
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/__init__.py
colossalai/fx/profiler/experimental/__init__.py
from .profiler import profile_function, profile_method, profile_module from .profiler_function import * from .profiler_module import * from .registry import meta_profiler_function, meta_profiler_module from .shard_utils import calculate_fwd_in, calculate_fwd_out, calculate_fwd_tmp
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/shard_utils.py
colossalai/fx/profiler/experimental/shard_utils.py
# for PyTorch 1.11 compatibility uses from torch.fx import Node from ..._compatibility import compatibility __all__ = ["calculate_fwd_in", "calculate_fwd_tmp", "calculate_fwd_out"] @compatibility(is_backward_compatible=True) def calculate_fwd_in(n: Node) -> bool: """A helper function to calculate `fwd_in` Args: n (Node): a node from the graph Returns: save_fwd_in (bool): the result of `save_fwd_in` """ return n.meta["save_fwd_in"] @compatibility(is_backward_compatible=True) def calculate_fwd_tmp(n: Node) -> int: """A helper function to calculate `fwd_tmp` Args: n (Node): a node from the graph Returns: fwd_tmp (int): the result of `fwd_tmp` """ return n.meta["fwd_mem_tmp"] @compatibility(is_backward_compatible=True) def calculate_fwd_out(n: Node) -> int: """A helper function to calculate `fwd_out` Args: n (Node): a node from the graph Returns: fwd_out (int): the result of `fwd_out` """ return n.meta["fwd_mem_out"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/torch_ops.py
colossalai/fx/profiler/experimental/profiler_function/torch_ops.py
import operator from functools import reduce from typing import Any, Optional, Tuple import torch from ..registry import meta_profiler_function @meta_profiler_function.register(torch.arange) @meta_profiler_function.register(torch.finfo) @meta_profiler_function.register(torch.permute) @meta_profiler_function.register(torch.Tensor.permute) @meta_profiler_function.register(torch.Tensor.repeat) @meta_profiler_function.register(torch.index_select) @meta_profiler_function.register(torch.Tensor.index_select) @meta_profiler_function.register(torch.squeeze) @meta_profiler_function.register(torch.Tensor.squeeze) @meta_profiler_function.register(torch.unsqueeze) @meta_profiler_function.register(torch.Tensor.unsqueeze) @meta_profiler_function.register(torch.cat) @meta_profiler_function.register(torch.concat) @meta_profiler_function.register(torch.repeat_interleave) @meta_profiler_function.register(torch.Tensor.repeat_interleave) @meta_profiler_function.register(torch.flatten) @meta_profiler_function.register(torch.Tensor.flatten) @meta_profiler_function.register(torch.roll) @meta_profiler_function.register(torch.full) @meta_profiler_function.register(torch.Tensor.cpu) @meta_profiler_function.register(torch.Tensor.cuda) @meta_profiler_function.register(torch._assert) def torch_zero_flops_op(*args, **kwargs) -> Tuple[int, int]: flops = 0 macs = 0 return flops, macs @meta_profiler_function.register(torch.where) def torch_where(condition: torch.Tensor, x: Any, y: Any) -> Tuple[int, int]: # torch.where returns the broadcasted tensor of condition, x, and y, # so hack it by using addition flops = condition.numel() macs = 0 return flops, macs @meta_profiler_function.register(torch.max) def torch_max( input: torch.Tensor, dim: int = None, keepdim: bool = False, *, out: Optional[torch.Tensor] = None ) -> Tuple[int, int]: macs = 0 assert out is None, "assigning value to out is not supported yet" if dim is not None: shape = list(input.shape) shape.pop(int(dim)) flops = reduce(operator.mul, shape), macs return flops, macs else: flops = input.numel() return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/arithmetic.py
colossalai/fx/profiler/experimental/profiler_function/arithmetic.py
# Copyright (c) Microsoft Corporation. # Licensed under the MIT License. import operator from functools import reduce from typing import Any, Optional, Tuple, Union import torch from ..registry import meta_profiler_function def _elementwise_flops_compute(input, other): # copied from https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/profiling/flops_profiler/profiler.py#L763 if not torch.is_tensor(input): if torch.is_tensor(other): return reduce(operator.mul, other.shape), 0 else: return 1, 0 elif not torch.is_tensor(other): return reduce(operator.mul, input.shape), 0 else: dim_input = len(input.shape) dim_other = len(other.shape) max_dim = max(dim_input, dim_other) final_shape = [] for i in range(max_dim): in_i = input.shape[i] if i < dim_input else 1 ot_i = other.shape[i] if i < dim_other else 1 if in_i > ot_i: final_shape.append(in_i) else: final_shape.append(ot_i) flops = reduce(operator.mul, final_shape) return flops, 0 @meta_profiler_function.register(torch.add) @meta_profiler_function.register(torch.eq) @meta_profiler_function.register(torch.sub) @meta_profiler_function.register(torch.mul) @meta_profiler_function.register(torch.floor_divide) @meta_profiler_function.register("add") # for built-in op + @meta_profiler_function.register("iadd") # for built-in op += @meta_profiler_function.register("eq") # for built-in op = @meta_profiler_function.register("sub") # for built-in op - @meta_profiler_function.register("isub") # for built-in op -= @meta_profiler_function.register("mul") # for built-in op * @meta_profiler_function.register("imul") # for built-in op *= @meta_profiler_function.register("floordiv") # for built-in op // @meta_profiler_function.register("ifloordiv") # for built-in op //= def torch_add_like_ops(input: Any, other: Any, *, out: Optional[torch.Tensor] = None) -> Tuple[int, int]: return _elementwise_flops_compute(input, other) @meta_profiler_function.register(torch.abs) def torch_elementwise_op(input: torch.Tensor, *, out: Optional[torch.Tensor] = None) -> Tuple[int, int]: flops = input.numel() macs = 0 return flops, macs @meta_profiler_function.register(torch.matmul) @meta_profiler_function.register("matmul") # for built-in op @ @meta_profiler_function.register(torch.Tensor.matmul) def torch_matmul(input: torch.Tensor, other: torch.Tensor, *, out: Optional[torch.Tensor] = None) -> Tuple[int, int]: macs = reduce(operator.mul, input.shape) * other.shape[-1] flops = 2 * macs return flops, macs @meta_profiler_function.register(torch.bmm) def torch_bmm(input: torch.Tensor, other: torch.Tensor, *, out: Optional[torch.Tensor] = None) -> Tuple[int, int]: macs = reduce(operator.mul, input.shape) * other.shape[-1] flops = 2 * macs return flops, macs @meta_profiler_function.register(torch.var_mean) def torch_var_mean( input: torch.Tensor, dim: Union[int, Tuple[int, ...]], unbiased: Optional[bool] = True, keepdim: Optional[bool] = False, *, out: Optional[torch.Tensor] = None, ) -> Tuple[int, int]: assert out is None, "saving to out is not supported yet" flops = input.numel() * 3 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/pooling.py
colossalai/fx/profiler/experimental/profiler_function/pooling.py
from typing import Tuple import torch from ..registry import meta_profiler_function @meta_profiler_function.register(torch.nn.functional.avg_pool1d) @meta_profiler_function.register(torch.nn.functional.avg_pool2d) @meta_profiler_function.register(torch.nn.functional.avg_pool3d) @meta_profiler_function.register(torch.nn.functional.max_pool1d) @meta_profiler_function.register(torch.nn.functional.max_pool2d) @meta_profiler_function.register(torch.nn.functional.max_pool3d) @meta_profiler_function.register(torch.nn.functional.adaptive_avg_pool1d) @meta_profiler_function.register(torch.nn.functional.adaptive_avg_pool2d) @meta_profiler_function.register(torch.nn.functional.adaptive_avg_pool3d) @meta_profiler_function.register(torch.nn.functional.adaptive_max_pool1d) @meta_profiler_function.register(torch.nn.functional.adaptive_max_pool2d) @meta_profiler_function.register(torch.nn.functional.adaptive_max_pool3d) def torch_nn_func_pooling(input: torch.Tensor, *args, **kwargs) -> Tuple[int, int]: # all pooling could be considered as going over each input element only once (https://stackoverflow.com/a/67301217) flops = input.numel() macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/activation_function.py
colossalai/fx/profiler/experimental/profiler_function/activation_function.py
from typing import Tuple import torch from ..registry import meta_profiler_function # TODO: different activation has different FLOPs count, currently unused. _multiplier = { torch.nn.functional.relu: 1, torch.nn.functional.prelu: 4, torch.nn.functional.sigmoid: 4, torch.nn.functional.tanh: 5, torch.nn.functional.leaky_relu: 3, torch.nn.functional.elu: 4, torch.nn.functional.relu6: 2, torch.nn.functional.gelu: 9, torch.nn.functional.hardswish: 5, torch.nn.functional.hardsigmoid: 4, } @meta_profiler_function.register(torch.nn.functional.leaky_relu) @meta_profiler_function.register(torch.nn.functional.elu) @meta_profiler_function.register(torch.nn.functional.gelu) @meta_profiler_function.register(torch.nn.functional.relu6) @meta_profiler_function.register(torch.nn.functional.prelu) @meta_profiler_function.register(torch.nn.functional.relu) @meta_profiler_function.register(torch.nn.functional.sigmoid) @meta_profiler_function.register(torch.nn.functional.tanh) @meta_profiler_function.register(torch.nn.functional.hardswish) @meta_profiler_function.register(torch.nn.functional.hardsigmoid) def torch_nn_func_non_linear_act(input: torch.Tensor, inplace: bool = False) -> Tuple[int, int]: flops = input.numel() macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/python_ops.py
colossalai/fx/profiler/experimental/profiler_function/python_ops.py
import operator from typing import Any, Tuple from ..registry import meta_profiler_function @meta_profiler_function.register(operator.getitem) def operator_getitem(a: Any, b: Any) -> Tuple[int, int]: flops = 0 macs = 0 return flops, macs @meta_profiler_function.register(getattr) def python_getattr(a: Any, b: Any) -> Tuple[int, int]: flops = 0 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/__init__.py
colossalai/fx/profiler/experimental/profiler_function/__init__.py
from .activation_function import * from .arithmetic import * from .embedding import * from .linear import * from .normalization import * from .pooling import * from .python_ops import * from .torch_ops import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/embedding.py
colossalai/fx/profiler/experimental/profiler_function/embedding.py
from typing import Optional import torch from ..registry import meta_profiler_function @meta_profiler_function.register(torch.nn.functional.embedding) def torch_nn_functional_embedding( input: torch.Tensor, weight: torch.Tensor, padding_idx: Optional[int] = None, max_norm: Optional[float] = None, norm_type: float = 2.0, scale_grad_by_freq: bool = False, sparse: bool = False, ) -> torch.Tensor: # F.embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6) flops = 0 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/normalization.py
colossalai/fx/profiler/experimental/profiler_function/normalization.py
from typing import List, Optional, Tuple import torch from ..registry import meta_profiler_function @meta_profiler_function.register(torch.nn.functional.instance_norm) def torch_nn_func_instancenorm( input: torch.Tensor, running_mean: Optional[torch.Tensor] = None, running_var: Optional[torch.Tensor] = None, weight: Optional[torch.Tensor] = None, bias: Optional[torch.Tensor] = None, use_input_stats: bool = True, momentum: float = 0.1, eps: float = 1e-5, ): has_affine = weight is not None flops = input.numel() * (5 if has_affine else 4) macs = 0 return flops, macs @meta_profiler_function.register(torch.nn.functional.group_norm) def torch_nn_func_groupnorm( input: torch.Tensor, num_groups: int, weight: Optional[torch.Tensor] = None, bias: Optional[torch.Tensor] = None, eps: float = 1e-5, ) -> Tuple[int, int]: has_affine = weight is not None flops = input.numel() * (5 if has_affine else 4) macs = 0 return flops, macs @meta_profiler_function.register(torch.nn.functional.layer_norm) def torch_nn_func_layernorm( input: torch.Tensor, normalized_shape: List[int], weight: Optional[torch.Tensor] = None, bias: Optional[torch.Tensor] = None, eps: float = 1e-5, ) -> Tuple[int, int]: has_affine = weight is not None flops = input.numel() * (5 if has_affine else 4) macs = 0 return flops, macs @meta_profiler_function.register(torch.nn.functional.batch_norm) def torch_nn_func_batchnorm( input: torch.Tensor, running_mean: Optional[torch.Tensor], running_var: Optional[torch.Tensor], weight: Optional[torch.Tensor] = None, bias: Optional[torch.Tensor] = None, training: bool = False, momentum: float = 0.1, eps: float = 1e-5, ) -> Tuple[int, int]: has_affine = weight is not None if training: flops = input.numel() * (2 if has_affine else 1) else: flops = input.numel() * (5 if has_affine else 4) macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_function/linear.py
colossalai/fx/profiler/experimental/profiler_function/linear.py
from typing import Tuple import torch from ..registry import meta_profiler_function @meta_profiler_function.register(torch.nn.functional.linear) def torch_nn_linear(input: torch.Tensor, weight: torch.Tensor, bias: torch.Tensor = None) -> Tuple[int, int]: out_features = weight.shape[0] macs = torch.numel(input) * out_features flops = 2 * macs if bias is not None: flops += bias.numel() return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/convolution.py
colossalai/fx/profiler/experimental/profiler_module/convolution.py
# Copyright (c) Microsoft Corporation. # Licensed under the MIT License. import math import operator from functools import reduce from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.Conv1d) def torch_nn_conv1d(self: torch.nn.Conv1d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html c_in, l_in = input.shape[-2:] c_out = self.out_channels l_out = math.floor( (l_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) result_shape = input.shape[:-2] + ( c_out, l_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce(operator.mul, result_shape) macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += num_elem return flops, macs @meta_profiler_module.register(torch.nn.Conv2d) def torch_nn_conv2d(self: torch.nn.Conv2d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv2d.html c_in, h_in, w_in = input.shape[-3:] c_out = self.out_channels h_out = math.floor( (h_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) w_out = math.floor( (w_in + 2 * self.padding[1] - self.dilation[1] * (self.kernel_size[1] - 1) - 1) / self.stride[1] + 1 ) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce(operator.mul, result_shape) macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += num_elem return flops, macs @meta_profiler_module.register(torch.nn.Conv3d) def torch_nn_conv3d(self: torch.nn.Conv3d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv3d.html c_in, d_in, h_in, w_in = input.shape[-4:] c_out = self.out_channels d_out = math.floor( (d_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) h_out = math.floor( (h_in + 2 * self.padding[1] - self.dilation[1] * (self.kernel_size[1] - 1) - 1) / self.stride[1] + 1 ) w_out = math.floor( (w_in + 2 * self.padding[2] - self.dilation[2] * (self.kernel_size[2] - 1) - 1) / self.stride[2] + 1 ) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce(operator.mul, result_shape) macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += num_elem return flops, macs @meta_profiler_module.register(torch.nn.ConvTranspose1d) def torch_nn_convtranspose1d(self: torch.nn.ConvTranspose1d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose1d.html c_in, l_in = input.shape[-2:] c_out = self.out_channels l_out = math.floor( (l_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) result_shape = input.shape[:-2] + ( c_out, l_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce( operator.mul, input.shape ) # see https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/profiling/flops_profiler/profiler.py#L604 macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += reduce(operator.mul, result_shape) return flops, macs @meta_profiler_module.register(torch.nn.ConvTranspose2d) def torch_nn_convtranspose2d(self: torch.nn.ConvTranspose2d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose2d.html c_in, h_in, w_in = input.shape[-3:] c_out = self.out_channels h_out = math.floor( (h_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) w_out = math.floor( (w_in - 1) * self.stride[1] - 2 * self.padding[1] + self.dilation[1] * (self.kernel_size[1] - 1) + self.output_padding[1] + 1 ) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce(operator.mul, input.shape) macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += reduce(operator.mul, result_shape) return flops, macs @meta_profiler_module.register(torch.nn.ConvTranspose3d) def torch_nn_convtranspose3d(self: torch.nn.ConvTranspose3d, input: torch.Tensor) -> Tuple[int, int]: # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose3d.html c_in, d_in, h_in, w_in = input.shape[-4:] c_out = self.out_channels d_out = math.floor( (d_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) h_out = math.floor( (h_in - 1) * self.stride[1] - 2 * self.padding[1] + self.dilation[1] * (self.kernel_size[1] - 1) + self.output_padding[1] + 1 ) w_out = math.floor( (w_in - 1) * self.stride[2] - 2 * self.padding[2] + self.dilation[2] * (self.kernel_size[2] - 1) + self.output_padding[2] + 1 ) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) macs_per_elem = reduce(operator.mul, self.kernel_size) * c_in // self.groups num_elem = reduce(operator.mul, input.shape) macs = macs_per_elem * num_elem flops = 2 * macs if self.bias is not None: flops += reduce(operator.mul, result_shape) return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/pooling.py
colossalai/fx/profiler/experimental/profiler_module/pooling.py
from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.AvgPool1d) @meta_profiler_module.register(torch.nn.AvgPool2d) @meta_profiler_module.register(torch.nn.AvgPool3d) @meta_profiler_module.register(torch.nn.MaxPool1d) @meta_profiler_module.register(torch.nn.MaxPool2d) @meta_profiler_module.register(torch.nn.MaxPool3d) @meta_profiler_module.register(torch.nn.AdaptiveAvgPool1d) @meta_profiler_module.register(torch.nn.AdaptiveMaxPool1d) @meta_profiler_module.register(torch.nn.AdaptiveAvgPool2d) @meta_profiler_module.register(torch.nn.AdaptiveMaxPool2d) @meta_profiler_module.register(torch.nn.AdaptiveAvgPool3d) @meta_profiler_module.register(torch.nn.AdaptiveMaxPool3d) def torch_nn_pooling(self: torch.nn.Module, input: torch.Tensor) -> Tuple[int, int]: # all pooling could be considered as going over each input element only once (https://stackoverflow.com/a/67301217) flops = input.numel() macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/torch_op.py
colossalai/fx/profiler/experimental/profiler_module/torch_op.py
from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.Flatten) def torch_nn_flatten(self: torch.nn.Flatten, input: torch.Tensor) -> Tuple[int, int]: flops = 0 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/rnn.py
colossalai/fx/profiler/experimental/profiler_module/rnn.py
import operator from functools import reduce from typing import Optional, Tuple import torch from ..registry import meta_profiler_module def _rnn_flops( flops: int, macs: int, module: torch.nn.RNNBase, w_ih: torch.Tensor, w_hh: torch.Tensor ) -> Tuple[int, int]: # copied from https://github.com/sovrasov/flops-counter.pytorch/blob/master/ptflops/pytorch_ops.py # matrix matrix mult ih state and internal state macs += reduce(operator.mul, w_ih.shape) flops += 2 * reduce(operator.mul, w_ih.shape) # matrix matrix mult hh state and internal state macs += reduce(operator.mul, w_hh.shape) flops += 2 * reduce(operator.mul, w_hh.shape) if isinstance(module, (torch.nn.RNN, torch.nn.RNNCell)): # add both operations flops += module.hidden_size elif isinstance(module, (torch.nn.GRU, torch.nn.GRUCell)): # hadamard of r flops += module.hidden_size # adding operations from both states flops += module.hidden_size * 3 # last two hadamard product and add flops += module.hidden_size * 3 elif isinstance(module, (torch.nn.LSTM, torch.nn.LSTMCell)): # adding operations from both states flops += module.hidden_size * 4 # two hadamard product and add for C state flops += module.hidden_size * 3 # final hadamard flops += module.hidden_size * 3 return flops, macs @meta_profiler_module.register(torch.nn.LSTM) @meta_profiler_module.register(torch.nn.GRU) @meta_profiler_module.register(torch.nn.RNN) def torch_nn_rnn(self: torch.nn.RNNBase, input: torch.Tensor, hx: Optional[torch.Tensor] = None) -> Tuple[int, int]: flops = 0 macs = 0 for i in range(self.num_layers): w_ih = self.__getattr__("weight_ih_l" + str(i)) w_hh = self.__getattr__("weight_hh_l" + str(i)) flops, macs = _rnn_flops(flops, macs, self, w_ih, w_hh) if self.bias: b_ih = self.__getattr__("bias_ih_l" + str(i)) b_hh = self.__getattr__("bias_hh_l" + str(i)) flops += reduce(operator.mul, b_ih) + reduce(operator.mul, b_hh) flops *= reduce(operator.mul, input.shape[:2]) macs *= reduce(operator.mul, input.shape[:2]) if self.bidirectional: flops *= 2 macs *= 2 return flops, macs @meta_profiler_module.register(torch.nn.LSTMCell) @meta_profiler_module.register(torch.nn.GRUCell) @meta_profiler_module.register(torch.nn.RNNCell) def torch_nn_rnn(self: torch.nn.RNNCellBase, input: torch.Tensor, hx: Optional[torch.Tensor] = None) -> Tuple[int, int]: flops = 0 macs = 0 w_ih = self.__getattr__("weight_ih_l") w_hh = self.__getattr__("weight_hh_l") flops, macs = _rnn_flops(flops, macs, self, w_ih, w_hh) if self.bias: b_ih = self.__getattr__("bias_ih_l") b_hh = self.__getattr__("bias_hh_l") flops += reduce(operator.mul, b_ih) + reduce(operator.mul, b_hh) flops *= input.shape[0] macs *= input.shape[0] return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/dropout.py
colossalai/fx/profiler/experimental/profiler_module/dropout.py
from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.Dropout) def torch_nn_dropout(self: torch.nn.Module, input: torch.Tensor) -> Tuple[int, int]: # nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6) flops = 0 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/activation_function.py
colossalai/fx/profiler/experimental/profiler_module/activation_function.py
from typing import Tuple import torch from ..registry import meta_profiler_module # TODO: different activation has different FLOPs count, currently unused. _multiplier = { torch.nn.ReLU: 1, torch.nn.PReLU: 4, torch.nn.Sigmoid: 4, torch.nn.Tanh: 5, torch.nn.LeakyReLU: 3, torch.nn.ELU: 4, torch.nn.ReLU6: 2, torch.nn.GELU: 9, torch.nn.Hardswish: 5, torch.nn.Hardsigmoid: 4, } @meta_profiler_module.register(torch.nn.ELU) @meta_profiler_module.register(torch.nn.LeakyReLU) @meta_profiler_module.register(torch.nn.ReLU) @meta_profiler_module.register(torch.nn.GELU) @meta_profiler_module.register(torch.nn.Sigmoid) @meta_profiler_module.register(torch.nn.Tanh) @meta_profiler_module.register(torch.nn.ReLU6) @meta_profiler_module.register(torch.nn.PReLU) @meta_profiler_module.register(torch.nn.Hardswish) @meta_profiler_module.register(torch.nn.Hardsigmoid) def torch_nn_non_linear_act(self: torch.nn.Module, input: torch.Tensor) -> Tuple[int, int]: flops = input.numel() macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/__init__.py
colossalai/fx/profiler/experimental/profiler_module/__init__.py
from .activation_function import * from .attention import * from .convolution import * from .dropout import * from .embedding import * from .linear import * from .normalization import * from .pooling import * from .rnn import * from .torch_op import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/attention.py
colossalai/fx/profiler/experimental/profiler_module/attention.py
from typing import Optional, Tuple import torch from ..registry import meta_profiler_module # TODO: This is hard to compute memory cost @meta_profiler_module.register(torch.nn.MultiheadAttention) def torch_nn_msa( self: torch.nn.MultiheadAttention, query: torch.Tensor, key: torch.Tensor, value: torch.Tensor, key_padding_mask: Optional[torch.Tensor] = None, need_weights: bool = True, attn_mask: Optional[torch.Tensor] = None, average_attn_weights: bool = True, ) -> Tuple[int, int]: if getattr(self, "batch_first", False): batch_size = query.shape[0] len_idx = 1 else: batch_size = query.shape[1] len_idx = 0 dim_idx = 2 qdim = query.shape[dim_idx] kdim = key.shape[dim_idx] vdim = value.shape[dim_idx] qlen = query.shape[len_idx] klen = key.shape[len_idx] vlen = value.shape[len_idx] num_heads = self.num_heads assert qdim == self.embed_dim if self.kdim is None: assert kdim == qdim if self.vdim is None: assert vdim == qdim flops = 0 macs = 0 # Q scaling flops += qlen * qdim # Initial projections flops += 2 * ((qlen * qdim * qdim) + (klen * kdim * kdim) + (vlen * vdim * vdim)) # QW # KW # VW macs += (qlen * qdim * qdim) + (klen * kdim * kdim) + (vlen * vdim * vdim) # QW # KW # VW if self.in_proj_bias is not None: flops += (qlen + klen + vlen) * qdim # attention heads: scale, matmul, softmax, matmul qk_head_dim = qdim // num_heads v_head_dim = vdim // num_heads head_flops = ( 2 * (qlen * klen * qk_head_dim) + (qlen * klen) + 2 * (qlen * klen * v_head_dim) # QK^T # softmax # AV ) head_macs = (qlen * klen * qk_head_dim) + 2 * (qlen * klen * v_head_dim) # QK^T # AV flops += num_heads * head_flops macs += num_heads * head_flops # final projection, bias is always enabled flops += qlen * vdim * (vdim + 1) flops *= batch_size macs *= batch_size return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/embedding.py
colossalai/fx/profiler/experimental/profiler_module/embedding.py
from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.Embedding) def torch_nn_embedding(self: torch.nn.Embedding, input: torch.Tensor) -> Tuple[int, int]: # nn.Embedding is a dictionary lookup, so technically it has 0 FLOPs. (https://discuss.pytorch.org/t/correct-way-to-calculate-flops-in-model/67198/6) flops = 0 macs = 0 return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/normalization.py
colossalai/fx/profiler/experimental/profiler_module/normalization.py
# Copyright (c) Microsoft Corporation. # Licensed under the MIT License. from typing import Tuple, Union import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.InstanceNorm1d) @meta_profiler_module.register(torch.nn.InstanceNorm2d) @meta_profiler_module.register(torch.nn.InstanceNorm3d) @meta_profiler_module.register(torch.nn.LayerNorm) @meta_profiler_module.register(torch.nn.GroupNorm) @meta_profiler_module.register(torch.nn.BatchNorm1d) @meta_profiler_module.register(torch.nn.BatchNorm2d) @meta_profiler_module.register(torch.nn.BatchNorm3d) def torch_nn_normalize( self: Union[ torch.nn.LayerNorm, torch.nn.GroupNorm, torch.nn.BatchNorm1d, torch.nn.BatchNorm2d, torch.nn.BatchNorm3d ], input: torch.Tensor, ) -> Tuple[int, int]: # adopted from https://github.com/microsoft/DeepSpeed/blob/master/deepspeed/profiling/flops_profiler/profiler.py#L615 has_affine = self.weight is not None if self.training: flops = input.numel() * (2 if has_affine else 1) else: flops = input.numel() * (5 if has_affine else 4) macs = 0 return flops, macs try: import apex meta_profiler_module.register(apex.normalization.FusedLayerNorm)(torch_nn_normalize) meta_profiler_module.register(apex.normalization.FusedRMSNorm)(torch_nn_normalize) meta_profiler_module.register(apex.normalization.MixedFusedLayerNorm)(torch_nn_normalize) meta_profiler_module.register(apex.normalization.MixedFusedRMSNorm)(torch_nn_normalize) except (ImportError, AttributeError): pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/profiler/experimental/profiler_module/linear.py
colossalai/fx/profiler/experimental/profiler_module/linear.py
from typing import Tuple import torch from ..registry import meta_profiler_module @meta_profiler_module.register(torch.nn.Linear) @meta_profiler_module.register(torch.nn.modules.linear.NonDynamicallyQuantizableLinear) def torch_nn_linear(self: torch.nn.Linear, input: torch.Tensor) -> Tuple[int, int]: out_features = self.weight.shape[0] macs = input.numel() * out_features flops = 2 * macs if self.bias is not None: flops += self.bias.numel() return flops, macs
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/_symbolic_trace.py
colossalai/fx/tracer/_symbolic_trace.py
from typing import Any, Callable, Dict, Optional, Union import torch from colossalai.fx import ColoGraphModule from colossalai.fx._compatibility import compatibility from .tracer import ColoTracer @compatibility(is_backward_compatible=True) def symbolic_trace( root: Union[torch.nn.Module, Callable[..., Any]], concrete_args: Optional[Dict[str, Any]] = None, meta_args: Optional[Dict[str, Any]] = None, trace_act_ckpt=False, ) -> ColoGraphModule: """ Symbolic tracing API Given an ``nn.Module`` or function instance ``root``, this function will return a ``ColoGraphModule`` constructed by recording operations seen while tracing through ``root``. With ``meta_args``, we can trace the model that are untraceable subject to control flow. If specified using ``meta_args`` only, the tracing can be done ahead of time. Note that ``meta_args`` are kwargs, which contains the key of the argument's names and the value of the argument's values. Uses: >>> model = ... # if this works >>> gm = symbolic_trace(model, concrete_args=concrete_args) # else try this >>> gm = symbolic_trace(model, concrete_args=concrete_args, meta_args={'x': torch.rand(1, 3, 224, 224, device='meta')}) Args: root (Union[torch.nn.Module, Callable[..., Any]]): Module or function to be traced and converted into a Graph representation. concrete_args (Optional[Dict[str, Any]], optional): Concrete arguments to be used for tracing. meta_args (Optional[Dict[str, Any]], optional): Inputs to be partially specialized, special for ``ColoTracer``. Defaults to None. Returns: ColoGraphModule: A ``ColoGraphModule`` created from the recorded operations from ``root``. Warnings: This API is still under development and can incur some bugs. Feel free to report any bugs to the Colossal-AI team. """ graph = ColoTracer(trace_act_ckpt=trace_act_ckpt).trace(root, concrete_args=concrete_args, meta_args=meta_args) name = root.__class__.__name__ if isinstance(root, torch.nn.Module) else root.__name__ return ColoGraphModule(root, graph, name)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/registry.py
colossalai/fx/tracer/registry.py
class PatchRegistry: def __init__(self, name): self.name = name self.store = {} def register(self, source): def wrapper(func): self.store[source] = func return func return wrapper def get(self, source): assert source in self.store target = self.store[source] return target def has(self, source): return source in self.store meta_patched_function = PatchRegistry(name="patched_functions_for_meta_execution") meta_patched_module = PatchRegistry(name="patched_modules_for_meta_execution") bias_addition_function = PatchRegistry(name="patched_function_for_bias_addition") bias_addition_module = PatchRegistry(name="patched_module_for_bias_addition") bias_addition_method = PatchRegistry(name="patched_method_for_bias_addition")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/tracer.py
colossalai/fx/tracer/tracer.py
#!/usr/bin/env python """ tracer.py: Implemented a tracer which supports control flow and user-defined meta arguments. The implementation is partly inspired HuggingFace's fx tracer """ import enum import functools import inspect import operator from contextlib import contextmanager from typing import Any, Dict, Optional import torch import torch.nn as nn from torch import Tensor from torch.fx import Node, Tracer from torch.fx.graph import Graph, magic_methods, reflectable_magic_methods from torch.fx.proxy import ParameterProxy, Proxy from ..proxy import ColoProxy from ._tracer_utils import compute_meta_data_for_functions_proxy, extract_meta, is_element_in_list from .bias_addition_patch import func_to_func_dict, method_to_func_dict, module_to_func_dict from .registry import ( bias_addition_function, bias_addition_method, bias_addition_module, meta_patched_function, meta_patched_module, ) __all__ = ["ColoTracer"] class TracerType(enum.Enum): DEFAULT = 1 META = 2 class ColoTracer(Tracer): """ ColoTracer is a symbolic tracer designed to support dynamic control flow by using meta tensors for the `colossalai.fx` module. This tracer is initialized in the same way as the original torch.fx.Tracer. Usage:: class Model(nn.Module): def __init__(self): super().__init__() self.linear1 = nn.Linear(10, 10) self.linear2 = nn.Linear(10, 10) def forward(self, x, y): x1 = self.linear1(x) y1 = self.linear2(y) if x1.dim() == 2: return x1 + y1 else: return x1 - y1 model = Model() tracer = ColoTracer() graph = tracer.trace(model, concrete_args={'y': torch.rand(4, 10)}, meta_args={'x': torch.rand(4, 10, device='meta')}) """ def __init__(self, trace_act_ckpt: bool = False, *args, **kwargs): super().__init__(*args, **kwargs) self.tracer_type = TracerType.META self.proxy_cls = ColoProxy # whether the tracer will record the usage of torch.utils.checkpoint self.trace_act_ckpt = trace_act_ckpt # whether the current tracing occurs within the activation checkpoint functions self.inside_torch_checkpoint_func = False self.act_ckpt_region_count = 0 # Feature flag for proxying accesses to buffer values proxy_buffer_attributes: bool = True _TORCH_METHODS_TO_PATCH = ["arange", "zeros", "ones", "full", "full_like", "eye", "empty", "tensor", "finfo"] def create_proxy(self, kind, target, args, kwargs, name=None, type_expr=None, proxy_factory_fn=None) -> ColoProxy: """ Create a proxy for different kinds of operations. """ if self.tracer_type == TracerType.DEFAULT: # since meta_args is not given # we just fall back to the original torch.fx.Tracer proxy = super().create_proxy(kind, target, args, kwargs, name, type_expr, proxy_factory_fn) return proxy # if graph is traced for auto parallelism module, some extra node will be added during # graph construction to deal with the compatibility between bias addition and all reduce. # if no extra manipulation is applied, we just pass the origin arguments to create_proxy function # to create node on computation graph origin_arguments = (kind, target, args, kwargs, name, type_expr, proxy_factory_fn) # dispatch the arguments generator depending on the kind and target in origin arguments. args_metas, _ = extract_meta(*args, **kwargs) handle = None if kind == "call_function": if bias_addition_function.has(target): if target == torch.nn.functional.linear: if "bias" in kwargs and kwargs["bias"] is not None: function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target)(self, target, args, kwargs, function_to_substitute) else: function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target)(self, target, args, kwargs, function_to_substitute) elif bias_addition_function.has(target.__name__): # use name for some builtin op like @ (matmul) function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target.__name__)(self, target, args, kwargs, function_to_substitute) elif kind == "call_method": method = getattr(args_metas[0].__class__, target) if bias_addition_method.has(method): function_to_substitute = method_to_func_dict[method] handle = bias_addition_method.get(method)(self, target, args, kwargs, function_to_substitute) elif kind == "call_module": if not hasattr(self, "orig_forward"): raise AttributeError(f"{self} does not have an attribute called orig_forward") self._disable_module_getattr = True try: mod = self.root.get_submodule(target) mod_type = type(mod) if bias_addition_module.has(mod_type) and mod.bias is not None: function_to_substitute = module_to_func_dict[mod_type] handle = bias_addition_module.get(mod_type)(self, target, args, kwargs, function_to_substitute) finally: self._disable_module_getattr = False if handle is not None: return handle.generate() # create nodes using patched arguments proxy = super().create_proxy(*origin_arguments) proxy: ColoProxy meta_out = self._meta_data_computing( kind, target, args, kwargs, ) proxy.meta_data = meta_out return proxy def _module_getattr(self, attr, attr_val, parameter_proxy_cache): if getattr(self, "_disable_module_getattr", False): return attr_val else: # return super()._module_getattr(attr, attr_val, parameter_proxy_cache) def maybe_get_proxy_for_attr(attr_val, collection_to_search, parameter_proxy_cache): for n, p in collection_to_search: if attr_val is p: if n not in parameter_proxy_cache: kwargs = {} if "proxy_factory_fn" in inspect.signature(self.create_proxy).parameters: kwargs["proxy_factory_fn"] = ( None if not self.param_shapes_constant else lambda node: ParameterProxy(self, node, n, attr_val) ) val_proxy = self.create_proxy("get_attr", n, (), {}, **kwargs) # type: ignore[arg-type] parameter_proxy_cache[n] = val_proxy return parameter_proxy_cache[n] return None if isinstance(attr_val, torch.nn.Parameter): maybe_parameter_proxy = maybe_get_proxy_for_attr( attr_val, self.root.named_parameters(), parameter_proxy_cache ) if maybe_parameter_proxy is not None: return maybe_parameter_proxy if self.proxy_buffer_attributes and isinstance(attr_val, torch.Tensor): maybe_buffer_proxy = maybe_get_proxy_for_attr( attr_val, self.root.named_buffers(), parameter_proxy_cache ) if maybe_buffer_proxy is not None: return maybe_buffer_proxy return attr_val def call_module(self, m, forward, args, kwargs): self.orig_forward = forward module_qualified_name = self.path_of_module(m) # a leaf module is the torch.nn.Module subclasses starting with `torch.nn` # which means customized modules are not leaf module by default # if a customized or third-party module like apex.normalization.FusedRMSNorm is patched, # we should treat it as leaf module as well if meta_patched_module.has(m.__class__) or self.is_leaf_module(m, module_qualified_name): return self.create_proxy("call_module", module_qualified_name, args, kwargs) else: return forward(*args, **kwargs) def proxy(self, node) -> Proxy: """ Returns a ColoProxy object. """ return self.proxy_cls(node, self) def _configure_tracer_type(self, tracer_type: TracerType): if tracer_type == TracerType.DEFAULT: self.proxy_cls = Proxy self.tracer_type = TracerType.DEFAULT elif tracer_type == TracerType.META: self.proxy_cls = ColoProxy self.tracer_type = TracerType.META else: raise ValueError(f"Unrecognized tracer type {tracer_type}") def _meta_data_computing(self, kind, target, args, kwargs): if kind == "placeholder" and target in self.meta_args and self.meta_args[target].is_meta: meta_out = self.meta_args[target] return meta_out if target in self.orig_torch_tensor_methods: # NOTE: tensor constructors in PyTorch define the `device` argument as # *kwargs-only*. That is why this works. If you add methods to # _TORCH_METHODS_TO_PATCH that do not define `device` as kwarg-only, # this will break and you will likely see issues where we cannot infer # the size of the output. if "device" in kwargs: kwargs["device"] = "meta" try: args_metas, kwargs_metas = extract_meta(*args, **kwargs) if kind == "call_function": # Our meta data will not record the nn.parameter.Parameter attribute。 # It works fine in most of the case, but it may cause some problems after # the bias addition manipulation. # Therefore, I need to record the nn.parameter.Parameter attribute for the operation # added by the bias addition manipulation following the get_attr node. convert_to_parameter = False if target in (torch.transpose, torch.reshape) and isinstance( args_metas[0], torch.nn.parameter.Parameter ): convert_to_parameter = True # fetch patched function if meta_patched_function.has(target): meta_target = meta_patched_function.get(target) elif meta_patched_function.has(target.__name__): # use name for some builtin op like @ (matmul) meta_target = meta_patched_function.get(target.__name__) else: meta_target = target meta_out = meta_target(*args_metas, **kwargs_metas) if isinstance(meta_out, torch.Tensor): meta_out = meta_out.to(device="meta") if convert_to_parameter: meta_out = torch.nn.Parameter(meta_out) elif kind == "call_method": # Our meta data will not record the nn.parameter.Parameter attribute。 # It works fine in most of the case, but it may cause some problems after # the bias addition manipulation. # Therefore, I need to record the nn.parameter.Parameter attribute for the operation # added by the bias addition manipulation following the get_attr node. convert_to_parameter = False if target in (torch.Tensor.view,) and isinstance(args_metas[0], torch.nn.parameter.Parameter): convert_to_parameter = True method = getattr(args_metas[0].__class__, target) # fetch patched method if meta_patched_function.has(method): meta_target = meta_patched_function.get(method) else: meta_target = method meta_out = meta_target(*args_metas, **kwargs_metas) if convert_to_parameter: meta_out = torch.nn.Parameter(meta_out) elif kind == "call_module": if not hasattr(self, "orig_forward"): raise AttributeError(f"{self} does not have an attribute called orig_forward") self._disable_module_getattr = True try: mod = self.root.get_submodule(target) mod_type = type(mod) if meta_patched_module.has(mod_type): meta_out = meta_patched_module.get(mod_type)(mod, *args_metas, **kwargs_metas) else: meta_out = self.orig_forward(*args_metas, **kwargs_metas) finally: self._disable_module_getattr = False elif kind == "get_attr": self._disable_module_getattr = True try: attr_itr = self.root atoms = target.split(".") for atom in atoms: attr_itr = getattr(attr_itr, atom) if isinstance(attr_itr, torch.nn.parameter.Parameter): meta_out = torch.nn.Parameter(attr_itr.to(device="meta")) elif isinstance(attr_itr, torch.Tensor): meta_out = attr_itr.to(device="meta") else: meta_out = attr_itr finally: self._disable_module_getattr = False else: return None except Exception as e: raise RuntimeError(f"Could not compute metadata for {kind} target {target}: {e}") return meta_out def trace( self, root: nn.Module, concrete_args: Optional[Dict[str, Tensor]] = None, meta_args: Optional[Dict[str, Tensor]] = None, ) -> Graph: """ Trace the forward computation graph using `torch.fx.Tracer`. This tracer enables data-dependent control flow. Args: root (nn.Module): a `nn.Module` object to trace the computation graph meta_args (Optional[Dict[str, Tensor]]): the meta tensor arguments used to trace the computation graph. These arguments are the sample data fed to the model during actual computation, but just converted to meta tensors. concrete_args (Optional[Dict[str, Tensor]]): the concrete arguments that should not be treated as Proxies. """ if meta_args is None: meta_args = {} if concrete_args is None: concrete_args = {} if len(meta_args) == 0: self._configure_tracer_type(TracerType.DEFAULT) else: self._configure_tracer_type(TracerType.META) # check concrete and meta args have valid names sig = inspect.signature(root.forward) sig_names = set(sig.parameters.keys()) meta_arg_names = set(meta_args.keys()) # update concrete args with default values non_meta_arg_names = sig_names - meta_arg_names for k, v in sig.parameters.items(): if k in non_meta_arg_names and k not in concrete_args and v.default is not inspect.Parameter.empty: concrete_args[k] = v.default # get non concrete arg names concrete_arg_names = set(concrete_args.keys()) non_concrete_arg_names = sig_names - concrete_arg_names def _check_arg_name_valid(names): success, element = is_element_in_list(names, sig_names) if not success: raise KeyError( f"argument {element} is not found in the signature of {root.__class__.__name__}'s forward function" ) _check_arg_name_valid(meta_arg_names) _check_arg_name_valid(concrete_arg_names) # assign as attributed for late reference def _check_kwargs(kwargs, should_be_meta: bool): for k, v in kwargs.items(): if not should_be_meta: assert ( not torch.is_tensor(v) or not v.is_meta ), f"Expected the {k} not to be a meta tensor, please check the args passed to the tracer" else: assert ( v.is_meta == should_be_meta ), f"Expected the is_meta attribute of {k} to be {should_be_meta}, but got {v.is_meta}, please check the args passed to the tracer" _check_kwargs(concrete_args, should_be_meta=False) _check_kwargs(meta_args, should_be_meta=True) self.concrete_args = concrete_args self.meta_args = meta_args self.patched_torch_tensor_methods = {} if self.tracer_type == TracerType.META: # wrap the torch tensor constructing methods so that they are captured in the graph self.patched_torch_tensor_methods = { target: wrap_tensor_constructor_method(getattr(torch, target)) for target in self._TORCH_METHODS_TO_PATCH } # patch these methods to replace their original use for name, (wrapper, orig) in self.patched_torch_tensor_methods.items(): setattr(torch, name, wrapper) # cache these methods so that we can detect whether a method call # should be patched during tracing self.orig_torch_tensor_methods = [val[1] for val in self.patched_torch_tensor_methods.values()] try: # to track the usage of torch.utils.checkpoint with self.trace_activation_checkpoint(enabled=self.trace_act_ckpt): self.graph = super().trace(root, concrete_args=concrete_args) finally: # recover the patched methods for name, (_, orig) in self.patched_torch_tensor_methods.items(): setattr(torch, name, orig) if self.tracer_type == TracerType.DEFAULT: return self.graph # This is necessary because concrete args are added as input to the traced module since # https://github.com/pytorch/pytorch/pull/55888. for node in self.graph.nodes: if node.op == "placeholder": # Removing default values for inputs as the forward pass will fail with them. if node.target in non_concrete_arg_names: node.args = () # Without this, torch.jit.script fails because the inputs type is Optional[torch.Tensor]. # It cannot infer on the attributes and methods the input should have, and fails. node.type = torch.Tensor # It is a concrete arg so it is not used and should be removed. else: if hasattr(torch.fx._symbolic_trace, "_assert_is_none"): # Newer versions of torch.fx emit an assert statement # for concrete arguments; delete those before we delete # the concrete arg. to_delete = [] for user in node.users: if user.target == torch.fx._symbolic_trace._assert_is_none: to_delete.append(user) for user in to_delete: self.graph.erase_node(user) self.graph.erase_node(node) # TODO: solves GraphModule creation. # Without this, return type annotation "Tuple" is causing code execution failure. if node.op == "output": node.type = None return self.graph @contextmanager def trace_activation_checkpoint(self, enabled: bool): if enabled: orig_ckpt_func = torch.utils.checkpoint.CheckpointFunction class PatchedCheckpointFunction(torch.autograd.Function): @staticmethod def forward(ctx, run_function, preserve_rng_state, *args): # signal that the current tracing occurs within activation checkpoint part self.inside_torch_checkpoint_func = True out = run_function(*args) self.inside_torch_checkpoint_func = False self.act_ckpt_region_count += 1 return out @staticmethod def backward(ctx: Any, *grad_outputs: Any) -> Any: raise NotImplementedError( "We do not implement the backward pass as we only trace the forward pass." ) # override the checkpoint function torch.utils.checkpoint.CheckpointFunction = PatchedCheckpointFunction yield if enabled: # recover the checkpoint function upon exit torch.utils.checkpoint.CheckpointFunction = orig_ckpt_func def create_node(self, *args, **kwargs) -> Node: node = super().create_node(*args, **kwargs) if self.inside_torch_checkpoint_func: # annotate the activation checkpoint module node.meta["activation_checkpoint"] = self.act_ckpt_region_count return node def wrap_tensor_constructor_method(target): def look_for_proxy(*args, **kwargs): # find in pos vars for arg in args: if isinstance(arg, Proxy): return arg if isinstance(arg, (tuple, list)): return look_for_proxy(*arg) # find in keyword vars for k, v in kwargs.items(): if isinstance(v, Proxy): return v if isinstance(v, (tuple, list)): return look_for_proxy(*v) return None @functools.wraps(target) def wrapper(*args, **kwargs): proxy = look_for_proxy(*args, **kwargs) if proxy is not None: # if the arg is a proxy, then need to record this function called on this proxy # e.g. torch.ones(size) where size is an input proxy colo_proxy = proxy.tracer.create_proxy("call_function", target, args, kwargs) if not isinstance(colo_proxy, ColoProxy): meta_out = compute_meta_data_for_functions_proxy(target, args, kwargs) colo_proxy = ColoProxy(proxy.node) colo_proxy.meta_data = meta_out return colo_proxy else: # this is called directly when the inputs do not contain proxy # e.g. torch.ones(4) where the input is static return target(*args, **kwargs) return wrapper, target # Patched magic methods for ColoProxy, then tracer could record the magic_method like __sub__, # and add meta_data attribute to the created proxy. for method in magic_methods: def _scope(method): def impl(*args, **kwargs): tracer = args[0].tracer target = getattr(operator, method) proxy = tracer.create_proxy("call_function", target, args, kwargs) if not isinstance(proxy, ColoProxy): meta_out = compute_meta_data_for_functions_proxy(target, args, kwargs) proxy = ColoProxy(proxy.node) proxy.meta_data = meta_out return proxy impl.__name__ = method as_magic = f'__{method.strip("_")}__' setattr(ColoProxy, as_magic, impl) _scope(method) def _define_reflectable(orig_method_name): method_name = f'__r{orig_method_name.strip("_")}__' def impl(self, rhs): target = getattr(operator, orig_method_name) proxy = self.tracer.create_proxy("call_function", target, (rhs, self), {}) if not isinstance(proxy, ColoProxy): meta_out = compute_meta_data_for_functions_proxy(target, *(rhs, self), {}) proxy = ColoProxy(proxy.node) proxy.meta_data = meta_out return proxy impl.__name__ = method_name impl.__qualname__ = method_name setattr(ColoProxy, method_name, impl) for orig_method_name in reflectable_magic_methods: _define_reflectable(orig_method_name)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/_meta_trace.py
colossalai/fx/tracer/_meta_trace.py
import torch from torch.fx import Graph, Node from torch.utils._pytree import tree_map def normalize_tuple(x): if not isinstance(x, tuple): return (x,) return x def is_autogradable(x): return isinstance(x, torch.Tensor) and x.is_floating_point() def meta_trace(module: torch.nn.Module, fake_device=None, *args, **kwargs) -> Graph: """Trace forward and backward graph with MetaTensor Args: module (torch.nn.Module): The target module for tracing. Returns: graph (torch.fx.Graph): The computation graph. Usage: >>> import torchvision.models as tm >>> model = tm.alexnet() >>> graph = meta_trace(model, torch.rand(1000, 3, 224, 224)) >>> graph.print_tabular() """ graph = Graph() namespace = graph._graph_namespace class MetaProxy(torch.Tensor): """ A wrapping tensor that hacks `torch.autograd` without patching more `torch.ops.aten` ops. """ _tensor: torch.Tensor _node: Node __slots__ = ["_tensor", "_node"] @staticmethod def __new__(cls, tensor, fake_device=None, placeholder=False, name=None): r = torch.Tensor._make_wrapper_subclass( cls, tensor.size(), strides=tensor.stride(), storage_offset=tensor.storage_offset(), dtype=tensor.dtype, layout=tensor.layout, device=fake_device if fake_device is not None else tensor.device, requires_grad=tensor.requires_grad, ) # deceive the frontend for aten selections r._tensor = tensor if placeholder: if name is None: name = "input" r._node = graph.create_node( "placeholder", "placeholder", (graph._root,), name=namespace.create_name(name, tensor) ) # ...the real tensor is held as an element on the tensor. if not r._tensor.is_meta: r._tensor = r._tensor.to(torch.device("meta")) return r @classmethod def __torch_dispatch__(cls, func, types, args=(), kwargs=None): def unwrap(x): nonlocal fake_device if isinstance(x, MetaProxy): fake_device = x.device x = x._tensor # assert not isinstance(x, MetaProxy) elif isinstance(x, torch.Tensor): fake_device = x.device x = x.to(torch.device("meta")) return x def get_node(x): if isinstance(x, torch.Tensor) and not hasattr(x, "_node"): x = MetaProxy(x, placeholder=True, name="weight") return x if not hasattr(x, "_node") else x._node args_node = tree_map(get_node, args) kwargs_node = tree_map(get_node, kwargs) node = graph.create_node("call_function", func, args_node, kwargs_node) if "device" in kwargs: fake_device = kwargs["device"] kwargs["device"] = torch.device("meta") args = tree_map(unwrap, args) kwargs = tree_map(unwrap, kwargs) # run aten for backend=CPU but actually on backend=Meta out = func(*args, **kwargs) # Now, we want to continue propagating this tensor, so we rewrap Tensors in # our custom tensor subclass def wrap(x): if isinstance(x, torch.Tensor): nonlocal fake_device if not x.is_meta: x = x.to(torch.device("meta")) return ( MetaProxy(x, fake_device=fake_device) if isinstance(x, torch.Tensor) and not hasattr(x, "_tensor") else x ) def set_node(x): x._node = node out = tree_map(wrap, out) tree_map(set_node, out) return out def wrap(x): return MetaProxy(x, fake_device=fake_device, placeholder=True) if isinstance(x, torch.Tensor) else x args = tree_map(wrap, args) kwargs = tree_map(wrap, kwargs) out = module(*args, **kwargs) for tensor in normalize_tuple(out): if is_autogradable(tensor) and tensor.requires_grad: grad = ( torch.empty_like(tensor._tensor, device=torch.device("meta")) if isinstance(tensor, MetaProxy) else torch.empty_like(tensor, device=torch.device("meta")) ) torch.autograd.backward( tensor, MetaProxy(grad, fake_device=tensor.device, placeholder=True), retain_graph=True ) return graph
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/_tracer_utils.py
colossalai/fx/tracer/_tracer_utils.py
from typing import Any, List, Union import torch from ..proxy import ColoProxy from .meta_patch import meta_patched_function __all__ = ["is_element_in_list", "extract_meta"] def is_element_in_list(elements: Union[List[Any], Any], list_: List[Any]): if isinstance(elements, (tuple, list, set)): for ele in elements: if ele not in list_: return False, ele else: if elements not in list_: return False, elements return True, None def extract_meta(*args, **kwargs): def _convert(val): if isinstance(val, ColoProxy): return val.meta_data elif isinstance(val, (list, tuple)): return type(val)([_convert(ele) for ele in val]) return val new_args = [_convert(val) for val in args] new_kwargs = {k: _convert(v) for k, v in kwargs.items()} return new_args, new_kwargs def compute_meta_data_for_functions_proxy(target, args, kwargs): args_metas, kwargs_metas = extract_meta(*args, **kwargs) # fetch patched function if meta_patched_function.has(target): meta_target = meta_patched_function.get(target) elif meta_patched_function.has(target.__name__): meta_target = meta_patched_function.get(target.__name__) else: meta_target = target meta_out = meta_target(*args_metas, **kwargs_metas) if isinstance(meta_out, torch.Tensor): meta_out = meta_out.to(device="meta") return meta_out
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/__init__.py
colossalai/fx/tracer/__init__.py
from colossalai.fx.tracer.meta_patch.patched_function.python_ops import operator_getitem from ._meta_trace import meta_trace from ._symbolic_trace import symbolic_trace from .tracer import ColoTracer
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/experimental.py
colossalai/fx/tracer/experimental.py
import functools import inspect import operator from contextlib import contextmanager from typing import Any, Callable, Dict, List, Optional, Set, Tuple, Union import torch from torch.fx import Graph, Node, Proxy, Tracer from torch.utils._pytree import tree_map from colossalai.fx import ColoGraphModule, compatibility, is_compatible_with_meta from colossalai.fx.tracer._tracer_utils import is_element_in_list from colossalai.fx.tracer.bias_addition_patch import func_to_func_dict, method_to_func_dict, module_to_func_dict from colossalai.fx.tracer.registry import ( bias_addition_function, bias_addition_method, bias_addition_module, meta_patched_function, meta_patched_module, ) if is_compatible_with_meta(): from colossalai.fx.profiler import MetaTensor Target = Union[Callable[..., Any], str] Argument = Optional[ Union[ Tuple[Any, ...], # actually Argument, but mypy can't represent recursive types List[Any], # actually Argument Dict[str, Any], # actually Argument slice, # Slice[Argument, Argument, Argument], but slice is not a templated type in typing "Node", ] ] _CScriptMethod = ["add", "mul", "sub", "div"] _TorchNewMethod = [ "arange", "zeros", "zeros_like", "ones", "ones_like", "full", "full_like", "empty", "empty_like", "eye", "tensor", "finfo", ] _TensorPropertyMethod = ["dtype", "shape", "device", "requires_grad", "grad", "grad_fn", "data"] def _truncate_suffix(s: str): import re return re.sub(r"_\d+$", "", s) def default_device(): return torch.device("cuda:0") if torch.cuda.is_available() else torch.device("cpu") @compatibility(is_backward_compatible=False) class ColoProxy(Proxy): def __init__(self, *args, data=None, **kwargs): super().__init__(*args, **kwargs) self._meta_data = data @property def meta_data(self): return self._meta_data @meta_data.setter def meta_data(self, args): wrap_fn = lambda x: MetaTensor(x) if isinstance(x, torch.Tensor) else x self._meta_data = tree_map(wrap_fn, args) @classmethod def __torch_function__(cls, orig_method, types, args=(), kwargs=None): proxy = cls.from_torch_proxy(super().__torch_function__(orig_method, types, args, kwargs)) unwrap_fn = lambda p: p.meta_data if isinstance(p, ColoProxy) else p kwargs = {} if kwargs is None else kwargs if proxy.meta_data is None: proxy.meta_data = orig_method(*tree_map(unwrap_fn, args), **tree_map(unwrap_fn, kwargs)) return proxy @classmethod def from_torch_proxy(cls, proxy: Proxy): return cls(proxy.node, proxy.tracer) def __repr__(self): return f"ColoProxy({self.node.name}, meta_data={self.meta_data})" def __len__(self): return len(self.meta_data) def __int__(self): return int(self.meta_data) def __index__(self): try: return int(self.meta_data) except: return torch.zeros(self.meta_data.shape, dtype=torch.bool).numpy().__index__() def __float__(self): return float(self.meta_data) def __bool__(self): return self.meta_data def __getattr__(self, k): return ColoAttribute(self, k, getattr(self._meta_data, k, None)) def __setitem__(self, key, value): proxy = self.tracer.create_proxy("call_function", operator.setitem, (self, key, value), {}) proxy.meta_data = self._meta_data return proxy def __contains__(self, key): if self.node.op == "placeholder": # this is used to handle like # if x in kwargs # we don't handle this case for now return False return super().__contains__(key) def __isinstancecheck__(self, type): return isinstance(self.meta_data, type) @property def shape(self): return self.meta_data.shape @property def ndim(self): return self.meta_data.ndim @property def device(self): proxy = self.tracer.create_proxy("call_function", getattr, (self, "device"), {}) proxy.meta_data = self.meta_data.device return proxy @property def dtype(self): proxy = self.tracer.create_proxy("call_function", getattr, (self, "dtype"), {}) proxy.meta_data = self.meta_data.dtype return proxy def to(self, *args, **kwargs): return self.tracer.create_proxy("call_method", "to", (self, *args), {**kwargs}) def cpu(self, *args, **kwargs): return self.tracer.create_proxy("call_method", "cpu", (self, *args), {**kwargs}) def cuda(self, *args, **kwargs): return self.tracer.create_proxy("call_method", "cuda", (self, *args), {**kwargs}) @compatibility(is_backward_compatible=False) class ColoAttribute(ColoProxy): def __init__(self, root, attr: str, data=None): self.root = root self.attr = attr self.tracer = root.tracer self._meta_data = data self._node: Optional[Node] = None @property def node(self): # the node for attributes is added lazily, since most will just be method calls # which do not rely on the getitem call if self._node is None: self._node = self.tracer.create_proxy("call_function", getattr, (self.root, self.attr), {}).node return self._node def __call__(self, *args, **kwargs): return self.tracer.create_proxy("call_method", self.attr, (self.root,) + args, kwargs) def __repr__(self): return f"ColoAttribute({self.node.name}, attr={self.attr})" @compatibility(is_backward_compatible=False) class ColoTracer(Tracer): def __init__(self, trace_act_ckpt: bool = False, *args, **kwargs): super().__init__(*args, **kwargs) self._disable_module_getattr = False self.proxy_buffer_attributes = True # whether the tracer will record the usage of torch.utils.checkpoint self.trace_act_ckpt = trace_act_ckpt # whether the current tracing occurs within the activation checkpoint functions self.inside_torch_checkpoint_func = False self.act_ckpt_region_count = 0 def proxy(self, node: Node) -> "ColoProxy": return ColoProxy(node, self) def create_proxy( self, kind: str, target: Target, args: Tuple[Any, ...], kwargs: Dict[str, Any], name: Optional[str] = None, type_expr: Optional[Any] = None, proxy_factory_fn: Callable[[Node], "Proxy"] = None, ): proxy: ColoProxy = super().create_proxy(kind, target, args, kwargs, name, type_expr, proxy_factory_fn) unwrap_fn = lambda p: p.meta_data if isinstance(p, ColoProxy) else p if kind == "placeholder": proxy.meta_data = ( self.meta_args[target] if target in self.meta_args else self.concrete_args.get(_truncate_suffix(target), None) ) elif kind == "get_attr": self._disable_module_getattr = True try: attr_itr = self.root atoms = target.split(".") for atom in atoms: attr_itr = getattr(attr_itr, atom) proxy.meta_data = attr_itr finally: self._disable_module_getattr = False elif kind == "call_function": proxy.meta_data = target(*tree_map(unwrap_fn, args), **tree_map(unwrap_fn, kwargs)) elif kind == "call_method": self._disable_module_getattr = True try: if target == "__call__": proxy.meta_data = unwrap_fn(args[0])(*tree_map(unwrap_fn, args[1:]), **tree_map(unwrap_fn, kwargs)) else: if target not in _TensorPropertyMethod: proxy._meta_data = getattr(unwrap_fn(args[0]), target)( *tree_map(unwrap_fn, args[1:]), **tree_map(unwrap_fn, kwargs) ) finally: self._disable_module_getattr = False elif kind == "call_module": mod = self.root.get_submodule(target) self._disable_module_getattr = True try: proxy.meta_data = mod.forward(*tree_map(unwrap_fn, args), **tree_map(unwrap_fn, kwargs)) finally: self._disable_module_getattr = False return proxy def create_node(self, *args, **kwargs) -> Node: node = super().create_node(*args, **kwargs) if self.inside_torch_checkpoint_func: # annotate the activation checkpoint module node.meta["activation_checkpoint"] = self.act_ckpt_region_count return node def trace( self, root: torch.nn.Module, concrete_args: Optional[Dict[str, torch.Tensor]] = None, meta_args: Optional[Dict[str, torch.Tensor]] = None, ) -> Graph: if meta_args is None: meta_args = {} if concrete_args is None: concrete_args = {} # check concrete and meta args have valid names sig = inspect.signature(root.forward) sig_names = set(sig.parameters.keys()) meta_arg_names = set(meta_args.keys()) # update concrete args with default values non_meta_arg_names = sig_names - meta_arg_names for k, v in sig.parameters.items(): if k in non_meta_arg_names and k not in concrete_args and v.default is not inspect.Parameter.empty: concrete_args[k] = v.default # get non concrete arg names concrete_arg_names = set(concrete_args.keys()) sig_names - concrete_arg_names def _check_arg_name_valid(names): success, element = is_element_in_list(names, sig_names) if not success: raise KeyError( f"argument {element} is not found in the signature of {root.__class__.__name__}'s forward function" ) _check_arg_name_valid(meta_arg_names) _check_arg_name_valid(concrete_arg_names) self.concrete_args = concrete_args self.meta_args = meta_args with _TorchTensorOverride(self), self.trace_activation_checkpoint(enabled=self.trace_act_ckpt): self.graph = super().trace(root, concrete_args=concrete_args) self.graph.lint() return self.graph @contextmanager def trace_activation_checkpoint(self, enabled: bool): if enabled: orig_ckpt_func = torch.utils.checkpoint.CheckpointFunction class PatchedCheckpointFunction(torch.autograd.Function): @staticmethod def forward(ctx, run_function, preserve_rng_state, *args): # signal that the current tracing occurs within activation checkpoint part self.inside_torch_checkpoint_func = True out = run_function(*args) self.inside_torch_checkpoint_func = False self.act_ckpt_region_count += 1 return out @staticmethod def backward(ctx: Any, *grad_outputs: Any) -> Any: raise NotImplementedError( "We do not implement the backward pass as we only trace the forward pass." ) # override the checkpoint function torch.utils.checkpoint.CheckpointFunction = PatchedCheckpointFunction yield if enabled: # recover the checkpoint function upon exit torch.utils.checkpoint.CheckpointFunction = orig_ckpt_func def _post_check(self, non_concrete_arg_names: Set[str]): # This is necessary because concrete args are added as input to the traced module since # https://github.com/pytorch/pytorch/pull/55888. for node in self.graph.nodes: if node.op == "placeholder": # Removing default values for inputs as the forward pass will fail with them. if node.target in non_concrete_arg_names: node.args = () # Without this, torch.jit.script fails because the inputs type is Optional[torch.Tensor]. # It cannot infer on the attributes and methods the input should have, and fails. node.type = torch.Tensor # It is a concrete arg so it is not used and should be removed. else: if hasattr(torch.fx._symbolic_trace, "_assert_is_none"): # Newer versions of torch.fx emit an assert statement # for concrete arguments; delete those before we delete # the concrete arg. to_delete = [] for user in node.users: if user.target == torch.fx._symbolic_trace._assert_is_none: to_delete.append(user) for user in to_delete: self.graph.erase_node(user) self.graph.erase_node(node) # TODO: solves GraphModule creation. # Without this, return type annotation "Tuple" is causing code execution failure. if node.op == "output": node.type = None self.graph.lint() def _module_getattr(self, attr, attr_val, parameter_proxy_cache): if getattr(self, "_disable_module_getattr", False): return attr_val def maybe_get_proxy_for_attr(attr_val, collection_to_search, parameter_proxy_cache): for n, p in collection_to_search: if attr_val is p: if n not in parameter_proxy_cache: kwargs = {} if "proxy_factory_fn" in inspect.signature(self.create_proxy).parameters: kwargs["proxy_factory_fn"] = ( None if not self.param_shapes_constant else lambda node: ColoProxy(self, node, n, attr_val) ) val_proxy = self.create_proxy("get_attr", n, (), {}, **kwargs) # type: ignore[arg-type] parameter_proxy_cache[n] = val_proxy return parameter_proxy_cache[n] return None if self.proxy_buffer_attributes and isinstance(attr_val, torch.Tensor): maybe_buffer_proxy = maybe_get_proxy_for_attr(attr_val, self.root.named_buffers(), parameter_proxy_cache) if maybe_buffer_proxy is not None: return maybe_buffer_proxy if isinstance(attr_val, torch.nn.Parameter): maybe_parameter_proxy = maybe_get_proxy_for_attr( attr_val, self.root.named_parameters(), parameter_proxy_cache ) if maybe_parameter_proxy is not None: return maybe_parameter_proxy return attr_val @compatibility(is_backward_compatible=True) def symbolic_trace( root: Union[torch.nn.Module, Callable[..., Any]], concrete_args: Optional[Dict[str, Any]] = None, meta_args: Optional[Dict[str, Any]] = None, trace_act_ckpt=False, ) -> ColoGraphModule: if is_compatible_with_meta(): if meta_args is not None: root.to(default_device()) wrap_fn = lambda x: MetaTensor(x, fake_device=default_device()) if isinstance(x, torch.Tensor) else x graph = ColoTracer(trace_act_ckpt=trace_act_ckpt).trace( root, concrete_args=concrete_args, meta_args=tree_map(wrap_fn, meta_args) ) root.cpu() else: graph = Tracer().trace(root, concrete_args=concrete_args) else: from .tracer import ColoTracer as OrigColoTracer graph = OrigColoTracer(trace_act_ckpt=trace_act_ckpt).trace( root, concrete_args=concrete_args, meta_args=meta_args ) name = root.__class__.__name__ if isinstance(root, torch.nn.Module) else root.__name__ return ColoGraphModule(root, graph, name) @compatibility(is_backward_compatible=False) class _TorchTensorOverride(object): def __init__(self, tracer: Tracer): self.overrides = {} self.tracer = tracer def __enter__(self): def wrap_tensor_method(target): @functools.wraps(target) def wrapper(*args, **kwargs): is_proxy = any(isinstance(p, ColoProxy) for p in args) | any( isinstance(p, ColoProxy) for p in kwargs.values() ) if is_proxy: # if the arg is a proxy, then need to record this function called on this proxy # e.g. torch.ones(size) where size is an input proxy self.tracer._disable_module_getattr = True try: proxy = self.tracer.create_proxy("call_function", target, args, kwargs) finally: self.tracer._disable_module_getattr = False return proxy else: return target(*args, **kwargs) return wrapper, target self.overrides = { target: wrap_tensor_method(getattr(torch, target)) for target in _TorchNewMethod if callable(getattr(torch, target)) } for name, (wrapper, orig) in self.overrides.items(): setattr(torch, name, wrapper) def __exit__(self, exc_type, exc_val, exc_tb): for name, (wrapper, orig) in self.overrides.items(): setattr(torch, name, orig) def meta_prop_pass( gm: ColoGraphModule, root: torch.nn.Module, meta_args: Optional[Dict[str, Any]] = None, concrete_args: Optional[Dict[str, torch.Tensor]] = None, ): if meta_args is None: meta_args = {} if concrete_args is None: concrete_args = {} # check concrete and meta args have valid names sig = inspect.signature(root.forward) sig_names = set(sig.parameters.keys()) meta_arg_names = set(meta_args.keys()) # update concrete args with default values non_meta_arg_names = sig_names - meta_arg_names for k, v in sig.parameters.items(): if k in non_meta_arg_names and k not in concrete_args and v.default is not inspect.Parameter.empty: concrete_args[k] = v.default for node in gm.graph.nodes: node._meta_data = _meta_data_computing( meta_args, concrete_args, root, node.op, node.target, node.args, node.kwargs ) def _meta_data_computing(meta_args, concrete_args, root, kind, target, args, kwargs): unwrap_fn = lambda n: n._meta_data if isinstance(n, Node) else n if kind == "placeholder": meta_out = meta_args[target] if target in meta_args else concrete_args.get(_truncate_suffix(target), None) elif kind == "get_attr": attr_itr = root atoms = target.split(".") for atom in atoms: attr_itr = getattr(attr_itr, atom) meta_out = attr_itr elif kind == "call_function": meta_out = target(*tree_map(unwrap_fn, args), **tree_map(unwrap_fn, kwargs)) elif kind == "call_method": if target == "__call__": meta_out = unwrap_fn(args[0])(*tree_map(unwrap_fn, args[1:]), **tree_map(unwrap_fn, kwargs)) else: if target not in _TensorPropertyMethod: meta_out = getattr(unwrap_fn(args[0]), target)( *tree_map(unwrap_fn, args[1:]), **tree_map(unwrap_fn, kwargs) ) elif kind == "call_module": mod = root.get_submodule(target) meta_out = mod.forward(*tree_map(unwrap_fn, args), **tree_map(unwrap_fn, kwargs)) else: meta_out = None return meta_out def _meta_data_computing_v0(meta_args, root, kind, target, args, kwargs): if kind == "placeholder" and target in meta_args and meta_args[target].is_meta: meta_out = meta_args[target] return meta_out if target in [getattr(torch, torch_func) for torch_func in _TorchNewMethod]: # NOTE: tensor constructors in PyTorch define the `device` argument as # *kwargs-only*. That is why this works. If you add methods to # _TORCH_METHODS_TO_PATCH that do not define `device` as kwarg-only, # this will break and you will likely see issues where we cannot infer # the size of the output. if "device" in kwargs: kwargs["device"] = "meta" try: unwrap_fn = lambda n: n._meta_data if isinstance(n, Node) else n args_metas = tree_map(unwrap_fn, args) kwargs_metas = tree_map(unwrap_fn, kwargs) if kind == "call_function": # fetch patched function if meta_patched_function.has(target): meta_target = meta_patched_function.get(target) elif meta_patched_function.has(target.__name__): # use name for some builtin op like @ (matmul) meta_target = meta_patched_function.get(target.__name__) else: meta_target = target meta_out = meta_target(*args_metas, **kwargs_metas) if isinstance(meta_out, torch.Tensor): meta_out = meta_out.to(device="meta") elif kind == "call_method": method = getattr(args_metas[0].__class__, target) # fetch patched method if meta_patched_function.has(method): meta_target = meta_patched_function.get(method) else: meta_target = method meta_out = meta_target(*args_metas, **kwargs_metas) elif kind == "call_module": mod = root.get_submodule(target) mod_type = type(mod) if meta_patched_module.has(mod_type): meta_out = meta_patched_module.get(mod_type)(mod, *args_metas, **kwargs_metas) else: meta_out = mod(*args_metas, **kwargs_metas) elif kind == "get_attr": attr_itr = root atoms = target.split(".") for atom in atoms: attr_itr = getattr(attr_itr, atom) if isinstance(attr_itr, torch.nn.parameter.Parameter): meta_out = torch.nn.Parameter(attr_itr.to(device="meta")) elif isinstance(attr_itr, torch.Tensor): meta_out = attr_itr.to(device="meta") else: meta_out = attr_itr else: return None except Exception as e: raise RuntimeError(f"Could not compute metadata for {kind} target {target}: {e}") return meta_out def bias_addition_pass(gm: ColoGraphModule, root_model: torch.nn.Module, meta_args: Optional[Dict[str, Any]] = None): result_graph = Graph() value_remap = {} unwrap_fn = lambda n: n._meta_data if isinstance(n, Node) else n for orig_node in gm.graph.nodes: assert hasattr(orig_node, "_meta_data") kind = orig_node.op target = orig_node.target args = orig_node.args kwargs = orig_node.kwargs args_metas = tree_map(unwrap_fn, args) tracer = ColoTracer() tracer.graph = Graph(tracer_cls=ColoTracer) tracer.root = root_model def wrap_fn(n): if isinstance(n, Node): proxy = ColoProxy(n, tracer) proxy.meta_data = n._meta_data return proxy return n args_proxy = tree_map(wrap_fn, args) kwargs_proxy = tree_map(wrap_fn, kwargs) handle = None if kind == "call_function": if bias_addition_function.has(target): if target == torch.nn.functional.linear: if "bias" in kwargs and kwargs["bias"] is not None: function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target)( tracer, target, args_proxy, kwargs_proxy, function_to_substitute ) else: function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target)( tracer, target, args_proxy, kwargs_proxy, function_to_substitute ) elif bias_addition_function.has(target.__name__): # use name for some builtin op like @ (matmul) function_to_substitute = func_to_func_dict[target] handle = bias_addition_function.get(target.__name__)( tracer, target, args_proxy, kwargs_proxy, function_to_substitute ) elif kind == "call_method": method = getattr(args_metas[0].__class__, target) if bias_addition_method.has(method): function_to_substitute = method_to_func_dict[method] handle = bias_addition_method.get(method)( tracer, target, args_proxy, kwargs_proxy, function_to_substitute ) elif kind == "call_module": # if not hasattr(self, "orig_forward"): # raise AttributeError(f"{self} does not have an attribute called orig_forward") mod = gm.get_submodule(target) mod_type = type(mod) if bias_addition_module.has(mod_type) and mod.bias is not None: function_to_substitute = module_to_func_dict[mod_type] handle = bias_addition_module.get(mod_type)( tracer, target, args_proxy, kwargs_proxy, function_to_substitute ) if handle is not None: handle.generate() for node_inserted in tracer.graph.nodes: value_remap[node_inserted] = result_graph.node_copy(node_inserted, lambda n: value_remap[n]) last_node = value_remap[node_inserted] value_remap[orig_node] = last_node else: value_remap[orig_node] = result_graph.node_copy(orig_node, lambda n: value_remap[n]) del tracer gm.graph = result_graph gm.recompile() meta_prop_pass(gm, root_model, meta_args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/__init__.py
colossalai/fx/tracer/meta_patch/__init__.py
from .patched_function import * from .patched_module import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/convolution.py
colossalai/fx/tracer/meta_patch/patched_function/convolution.py
import collections import math from itertools import repeat import torch from ...registry import meta_patched_function def _ntuple(n, name="parse"): def parse(x): if isinstance(x, collections.abc.Iterable): return tuple(x) return tuple(repeat(x, n)) parse.__name__ = name return parse _single = _ntuple(1, "_single") _pair = _ntuple(2, "_pair") _triple = _ntuple(3, "_triple") def _extract_kwargs(kwargs): if "stride" in kwargs: stride = kwargs["stride"] else: stride = 1 # TODO: process str type padding if "padding" in kwargs: padding = kwargs["padding"] else: padding = 0 if "dilation" in kwargs: dilation = kwargs["dilation"] else: dilation = 1 if "output_padding" in kwargs: output_padding = kwargs["output_padding"] else: output_padding = 0 return stride, padding, dilation, output_padding @meta_patched_function.register(torch.nn.functional.conv1d) def torch_nn_functional_conv1d(input, weight, **kwargs): stride, padding, dilation, _ = _extract_kwargs(kwargs) stride = _single(stride) padding = _single(padding) dilation = _single(dilation) kernel_size = weight.shape[2:] l_in = input.shape[-1] c_out = weight.shape[0] l_out = math.floor((l_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) result_shape = input.shape[:-2] + ( c_out, l_out, ) return torch.empty(result_shape, device="meta") @meta_patched_function.register(torch.nn.functional.conv2d) def torch_nn_functional_conv2d(input, weight, **kwargs): stride, padding, dilation, _ = _extract_kwargs(kwargs) stride = _pair(stride) padding = _pair(padding) dilation = _pair(dilation) kernel_size = weight.shape[2:] h_in, w_in = input.shape[-2:] c_out = weight.shape[0] h_out = math.floor((h_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) w_out = math.floor((w_in + 2 * padding[1] - dilation[1] * (kernel_size[1] - 1) - 1) / stride[1] + 1) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_function.register(torch.nn.functional.conv3d) def torch_nn_functional_conv3d(input, weight, **kwargs): stride, padding, dilation, _ = _extract_kwargs(kwargs) stride = _triple(stride) padding = _triple(padding) dilation = _triple(dilation) kernel_size = weight.shape[2:] d_in, h_in, w_in = input.shape[-3:] c_out = weight.shape[0] d_out = math.floor((d_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) h_out = math.floor((h_in + 2 * padding[1] - dilation[1] * (kernel_size[1] - 1) - 1) / stride[1] + 1) w_out = math.floor((w_in + 2 * padding[2] - dilation[2] * (kernel_size[2] - 1) - 1) / stride[2] + 1) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_function.register(torch.nn.functional.conv_transpose1d) def torch_nn_functional_convtranspose1d(input, weight, **kwargs): stride, padding, dilation, output_padding = _extract_kwargs(kwargs) stride = _single(stride) padding = _single(padding) dilation = _single(dilation) output_padding = _single(output_padding) kernel_size = weight.shape[2:] l_in = input.shape[-1] c_out = weight.shape[1] l_out = math.floor( (l_in - 1) * stride[0] - 2 * padding[0] + dilation[0] * (kernel_size[0] - 1) + output_padding[0] + 1 ) result_shape = input.shape[:-2] + ( c_out, l_out, ) return torch.empty(result_shape, device="meta") @meta_patched_function.register(torch.nn.functional.conv_transpose2d) def torch_nn_functional_convtranspose2d(input, weight, **kwargs): stride, padding, dilation, output_padding = _extract_kwargs(kwargs) stride = _pair(stride) padding = _pair(padding) dilation = _pair(dilation) output_padding = _pair(output_padding) kernel_size = weight.shape[2:] h_in, w_in = input.shape[-2:] c_out = weight.shape[1] h_out = math.floor( (h_in - 1) * stride[0] - 2 * padding[0] + dilation[0] * (kernel_size[0] - 1) + output_padding[0] + 1 ) w_out = math.floor( (w_in - 1) * stride[1] - 2 * padding[1] + dilation[1] * (kernel_size[1] - 1) + output_padding[1] + 1 ) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_function.register(torch.nn.functional.conv_transpose3d) def torch_nn_functional_convtranspose3d(input, weight, **kwargs): stride, padding, dilation, output_padding = _extract_kwargs(kwargs) stride = _triple(stride) padding = _triple(padding) dilation = _triple(dilation) output_padding = _triple(output_padding) kernel_size = weight.shape[2:] d_in, h_in, w_in = input.shape[-3:] c_out = weight.shape[1] d_out = math.floor( (d_in - 1) * stride[0] - 2 * padding[0] + dilation[0] * (kernel_size[0] - 1) + output_padding[0] + 1 ) h_out = math.floor( (h_in - 1) * stride[1] - 2 * padding[1] + dilation[1] * (kernel_size[1] - 1) + output_padding[1] + 1 ) w_out = math.floor( (w_in - 1) * stride[2] - 2 * padding[2] + dilation[2] * (kernel_size[2] - 1) + output_padding[2] + 1 ) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/torch_ops.py
colossalai/fx/tracer/meta_patch/patched_function/torch_ops.py
import torch from ...registry import meta_patched_function @meta_patched_function.register(torch.arange) def torch_arange(*args, **kwargs): n = len(args) step = 1 if n == 1: start = 0 end = args[0] elif n == 2: start, end = args else: start, end, step = args if isinstance(start, float): start = int(start) if isinstance(end, float): start = int(end) if isinstance(step, float): step = int(step) step = kwargs.get("step", step) dtype = kwargs.get("dtype") return torch.empty((end - start) // step, dtype=dtype, device="meta") @meta_patched_function.register(torch.finfo) def torch_finfo(*args): return torch.finfo(*args) @meta_patched_function.register(torch.where) def torch_where(condition, x, y): # torch.where returns the broadcasted tensor of condition, x, and y, # so hack it by using addition return condition.to(device="meta") + x.to(device="meta") + y.to(device="meta") @meta_patched_function.register(torch.Tensor.repeat) def torch_tensor_repeat(self, *sizes): shape = list(self.shape) for i, x in enumerate(sizes): shape[i] *= x return torch.empty(shape, device="meta") @meta_patched_function.register(torch.index_select) def torch_index_select(input, dim, index, *, out=None): shape = list(input.shape) shape[dim] = len(index) return torch.empty(*shape, device="meta") @meta_patched_function.register(torch.Tensor.index_select) def torch_tensor_index_select(self, dim, index): return torch_index_select(self, dim, index) @meta_patched_function.register(torch.squeeze) def torch_squeeze(input, dim=None): shape = list(input.shape) if dim is not None: if dim < 0: dim = input.dim() + dim if shape[dim] == 1: shape.pop(dim) else: new_shape = [] for dim_value in shape: if dim_value == 1: continue new_shape.append(dim_value) shape = new_shape return torch.empty(shape, device="meta") @meta_patched_function.register(torch.Tensor.squeeze) def torch_tensor_squeeze(self, dim=None): return torch_squeeze(self, dim) @meta_patched_function.register(torch.unsqueeze) def torch_unsqueeze(input, dim): shape = list(input.shape) if dim < 0: dim = input.dim() + 1 + dim shape.insert(dim, 1) return torch.empty(shape, device="meta") @meta_patched_function.register(torch.Tensor.unsqueeze) def torch_tensor_unsqueeze(self, dim): return torch_unsqueeze(self, dim) @meta_patched_function.register(torch.cat) def torch_cat(tensors, dim=None, axis=None, *, out=None): if dim is None and axis is None: dim = 0 if dim is None and axis is not None: dim = axis if dim < 0: dim = tensors[0].dim() + dim shapes = [t.shape for t in tensors] shape = list(shapes[0]) concatenated_dim = sum(shape[dim] for shape in shapes) final_shape = shape[:dim] + [concatenated_dim] + shape[dim + 1 :] return torch.empty(final_shape, device="meta") @meta_patched_function.register(torch.repeat_interleave) def torch_repeat_interleave(input, repeats, dim=None, output_size=None): assert isinstance(repeats, int) or isinstance( repeats, torch.Tensor ), "Argument 'repeats' should be of type 'torch.Tensor' or 'int'" shape = list(input.shape) if dim is not None else [input.numel()] dim = dim if dim is not None else 0 dim = input.dim() + dim if dim < 0 else dim if isinstance(repeats, int): shape[dim] = shape[dim] * repeats elif isinstance(repeats, torch.Tensor): shape[dim] = repeats.sum() return torch.empty(shape, device="meta") @meta_patched_function.register(torch.Tensor.repeat_interleave) def torch_tensor_repeat_interleave(self, repeats, dim=None, *, output_size=None): return torch_repeat_interleave(self, repeats, dim, output_size) @meta_patched_function.register(torch.roll) def torch_roll(input, shifts, dims=None): return torch.empty(input.shape, device="meta") @meta_patched_function.register(torch.full) def torch_full(size, fill_value, *, out=None, dtype=None, layout=torch.strided, device=None, requires_grad=False): assert out is None, "assigning result to out is not supported yet" return torch.empty(size, device="meta", dtype=dtype, layout=layout, requires_grad=requires_grad) @meta_patched_function.register(torch.max) def torch_max(input, dim=None, keepdim=False, *, out=None): assert out is None, "assigning value to out is not supported yet" if dim is not None: if isinstance(dim, int): shape = list(input.shape) shape.pop(dim) if keepdim: shape.insert(dim, 1) return torch.empty(shape, device="meta", dtype=input.dtype), torch.empty( shape, device="meta", dtype=input.dtype ) elif isinstance(dim, torch.Tensor): # when dim is a 0D or 1D tensor, it will maintain the same shape num_dims = dim.dim() if num_dims in [0, 1]: return torch.empty_like(input, device="meta") else: raise ValueError(f"Expected dim to a 0D or 1D tensor but got {num_dims} dimensions") else: return torch.empty([], device="meta", dtype=input.dtype) @meta_patched_function.register(torch.Tensor.cpu) def torch_tensor_cpu(input): return input.clone() @meta_patched_function.register(torch.Tensor.cuda) def torch_tensor_cuda(input, *args, **kwargs): return input.clone()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/arithmetic.py
colossalai/fx/tracer/meta_patch/patched_function/arithmetic.py
import torch from ...registry import meta_patched_function @meta_patched_function.register(torch.matmul) @meta_patched_function.register("matmul") # for built-in op @ def torch_matmul(input, other, *, out=None): # copied from huggingface.utils.fx d1 = input.dim() d2 = other.dim() shape = None if d1 == 1 and d2 == 1: shape = None elif d1 == 2 and d2 == 2: shape = (input.size(0), other.size(1)) elif d1 == 1 and d2 == 2: shape = (other.size(1),) elif d1 == 2 and d2 == 1: shape = (input.size(0),) else: max_length = max(input.dim(), other.dim()) shape1 = list(input.shape) shape2 = list(other.shape) if d1 == 1: shape1 = [1] + shape1 if d2 == 1: shape2.append(1) shape1 = [-1] * (max_length - d1) + list(input.shape) shape2 = [-1] * (max_length - d2) + list(other.shape) shape = [] for i in range(max_length): shape.append(max(shape1[i], shape2[i])) shape[-2] = shape1[-2] shape[-1] = shape2[-1] if d1 == 1: shape.pop(-2) if d2 == 1: shape.pop(-1) if shape is None: return torch.tensor(0.0, device="meta") return torch.empty(*shape, device="meta") @meta_patched_function.register(torch.abs) def torch_abs(input, *, out=None): assert out is None, "out is not supported yet" return torch.empty(input.shape, device="meta") @meta_patched_function.register(torch.bmm) def torch_bmm(input, mat2, *, out=None): if out is not None: raise ValueError("Don't support in-place abs for MetaTensor analysis") batch_size, n, m = input.shape _, _, p = mat2.shape return torch.empty(batch_size, n, p, device="meta") @meta_patched_function.register(torch.nn.functional.linear) def torch_linear(input, mat2, bias=None, *, out=None): if out is not None: raise ValueError("Don't support in-place abs for MetaTensor analysis") output_shape = list(input.shape) output_feature = list(mat2.shape)[0] output_shape[-1] = output_feature return torch.empty(*output_shape, device="meta") @meta_patched_function.register(torch.addbmm) @meta_patched_function.register(torch.Tensor.addbmm) def torch_addbmm(input, mat1, mat2, *, beta=1, alpha=1, out=None): if out is not None: raise ValueError("Don't support in-place abs for MetaTensor analysis") _, n, _ = mat1.shape _, _, p = mat2.shape return torch.empty(n, p, device="meta") @meta_patched_function.register(torch.addmm) @meta_patched_function.register(torch.Tensor.addmm) def torch_addmm(input, mat1, mat2, *, beta=1, alpha=1, out=None): if out is not None: raise ValueError("Don't support in-place abs for MetaTensor analysis") n, _ = mat1.shape _, p = mat2.shape return torch.empty(n, p, device="meta") @meta_patched_function.register(torch.var_mean) def torch_var_mean(input, dim, unbiased=True, keepdim=False, *, out=None): assert out is None, "saving to out is not supported yet" var = torch.empty(1).squeeze(0).to("meta") mean = torch.empty(1).squeeze(0).to("meta") return var, mean
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/activation_function.py
colossalai/fx/tracer/meta_patch/patched_function/activation_function.py
import torch from ...registry import meta_patched_function @meta_patched_function.register(torch.nn.functional.relu) def torch_nn_func_relu(input, inplace=False): return torch.empty(input.shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/python_ops.py
colossalai/fx/tracer/meta_patch/patched_function/python_ops.py
import operator import torch from colossalai.fx.proxy import ColoProxy from ...registry import meta_patched_function @meta_patched_function.register(operator.getitem) def operator_getitem(a, b): # copied from huggingface.utils.fx def to_concrete(t): if isinstance(t, torch.Tensor): concrete = torch.ones_like(t, device="cpu") if concrete.dtype in [torch.float16, torch.float32, torch.float64, torch.int32]: concrete = concrete.to(torch.int64) return concrete return t def _slice_convert(slice_obj): attrs = {"start": slice_obj.start, "stop": slice_obj.stop, "step": slice_obj.step} new_attrs = _slice_attr_convert(attrs) attr_dict_to_tuple = (new_attrs["start"], new_attrs["stop"], new_attrs["step"]) return slice(*attr_dict_to_tuple) def _slice_attr_convert(attrs): new_attrs = {} for key, value in attrs.items(): if isinstance(value, ColoProxy): new_attrs[key] = value.meta_data else: new_attrs[key] = value return new_attrs if isinstance(b, tuple): b = list(b) for index, element in enumerate(b): if isinstance(element, slice): b[index] = _slice_convert(element) b = tuple(b) elif isinstance(b, slice): b = _slice_convert(b) if isinstance(a, torch.Tensor): # TODO: infer shape without performing the computation. if isinstance(b, tuple): b = tuple(map(to_concrete, b)) else: b = to_concrete(b) return operator.getitem(torch.empty_like(a, device="cpu"), b).to("meta") if isinstance(a, ColoProxy): # TODO: infer shape without performing the computation. if isinstance(b, tuple): b = tuple(map(to_concrete, b)) else: b = to_concrete(b) return operator.getitem(torch.empty_like(a.meta_data, device="cpu"), b).to("meta") return operator.getitem(a, b)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/__init__.py
colossalai/fx/tracer/meta_patch/patched_function/__init__.py
from .activation_function import * from .arithmetic import * from .convolution import * from .embedding import * from .normalization import * from .torch_ops import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/embedding.py
colossalai/fx/tracer/meta_patch/patched_function/embedding.py
import torch from ...registry import meta_patched_function @meta_patched_function.register(torch.nn.functional.embedding) def torch_nn_functional_embedding( input, weight, padding_idx=None, max_norm=None, norm_type=2.0, scale_grad_by_freq=False, sparse=False ): return torch.empty(*input.shape, weight.shape[-1], device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_function/normalization.py
colossalai/fx/tracer/meta_patch/patched_function/normalization.py
import torch from ...registry import meta_patched_function @meta_patched_function.register(torch.nn.functional.layer_norm) def torch_nn_func_layernorm(input, normalized_shape, weight=None, bias=None, eps=1e-05): return torch.empty(input.shape, device="meta") @meta_patched_function.register(torch.nn.functional.batch_norm) def torch_nn_func_batchnorm( input, running_mean, running_var, weight=None, bias=None, training=False, momentum=0.1, eps=1e-05 ): return torch.empty(input.shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/convolution.py
colossalai/fx/tracer/meta_patch/patched_module/convolution.py
import math import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.Conv1d) def torch_nn_conv1d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html#torch.nn.Conv1d l_in = input.shape[-1] c_out = self.out_channels l_out = math.floor( (l_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) result_shape = input.shape[:-2] + ( c_out, l_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.Conv2d) def torch_nn_conv2d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html#torch.nn.Conv2d h_in, w_in = input.shape[-2:] c_out = self.out_channels h_out = math.floor( (h_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) w_out = math.floor( (w_in + 2 * self.padding[1] - self.dilation[1] * (self.kernel_size[1] - 1) - 1) / self.stride[1] + 1 ) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.Conv3d) def torch_nn_conv3d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.Conv1d.html#torch.nn.Conv3d d_in, h_in, w_in = input.shape[-3:] c_out = self.out_channels d_out = math.floor( (d_in + 2 * self.padding[0] - self.dilation[0] * (self.kernel_size[0] - 1) - 1) / self.stride[0] + 1 ) h_out = math.floor( (h_in + 2 * self.padding[1] - self.dilation[1] * (self.kernel_size[1] - 1) - 1) / self.stride[1] + 1 ) w_out = math.floor( (w_in + 2 * self.padding[2] - self.dilation[2] * (self.kernel_size[2] - 1) - 1) / self.stride[2] + 1 ) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.ConvTranspose1d) def torch_nn_convtranspose1d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose1d.html l_in = input.shape[-1] c_out = self.out_channels l_out = math.floor( (l_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) result_shape = input.shape[:-2] + ( c_out, l_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.ConvTranspose2d) def torch_nn_convtranspose2d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose2d.html h_in, w_in = input.shape[-2:] c_out = self.out_channels h_out = math.floor( (h_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) w_out = math.floor( (w_in - 1) * self.stride[1] - 2 * self.padding[1] + self.dilation[1] * (self.kernel_size[1] - 1) + self.output_padding[1] + 1 ) result_shape = input.shape[:-3] + ( c_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.ConvTranspose3d) def torch_nn_convtranspose3d(self, input): # the output shape is calculated using the formula stated # at https://pytorch.org/docs/stable/generated/torch.nn.ConvTranspose3d.html d_in, h_in, w_in = input.shape[-3:] c_out = self.out_channels d_out = math.floor( (d_in - 1) * self.stride[0] - 2 * self.padding[0] + self.dilation[0] * (self.kernel_size[0] - 1) + self.output_padding[0] + 1 ) h_out = math.floor( (h_in - 1) * self.stride[1] - 2 * self.padding[1] + self.dilation[1] * (self.kernel_size[1] - 1) + self.output_padding[1] + 1 ) w_out = math.floor( (w_in - 1) * self.stride[2] - 2 * self.padding[2] + self.dilation[2] * (self.kernel_size[2] - 1) + self.output_padding[2] + 1 ) result_shape = input.shape[:-4] + ( c_out, d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/pooling.py
colossalai/fx/tracer/meta_patch/patched_module/pooling.py
import math import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.AvgPool1d) def torch_nn_avgpool1d(self, input): num_dim = input.dim() assert num_dim in [2, 3], f"expected the input to have 2 or 3 dimensions, but got {num_dim} dimensions" l_in = input.shape[-1] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 1 else: return item padding = _convert_int_to_list(self.padding) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) l_out = math.floor((l_in + 2 * padding[0] - kernel_size[0]) / stride[0] + 1) result_shape = tuple(input.shape[:-1]) + (l_out,) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.AvgPool2d) def torch_nn_avgpool2d(self, input): num_dim = input.dim() assert num_dim in [3, 4], f"expected the input to have 3 or 4 dimensions, but got {num_dim} dimensions" h_in, w_in = input.shape[-2:] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 2 else: return item padding = _convert_int_to_list(self.padding) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) h_out = math.floor((h_in + 2 * padding[0] - kernel_size[0]) / stride[0] + 1) w_out = math.floor((w_in + 2 * padding[1] - kernel_size[1]) / stride[1] + 1) result_shape = tuple(input.shape[:-2]) + ( h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.AvgPool3d) def torch_nn_avgpool3d(self, input): num_dim = input.dim() assert num_dim in [4, 5], f"expected the input to have 4 or 5 dimensions, but got {num_dim} dimensions" d_in, h_in, w_in = input.shape[-3:] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 3 else: return item padding = _convert_int_to_list(self.padding) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) d_out = math.floor((d_in + 2 * padding[0] - kernel_size[0]) / stride[0] + 1) h_out = math.floor((h_in + 2 * padding[1] - kernel_size[1]) / stride[1] + 1) w_out = math.floor((w_in + 2 * padding[2] - kernel_size[2]) / stride[2] + 1) result_shape = tuple(input.shape[:-3]) + ( d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.MaxPool1d) def torch_nn_maxpool1d(self, input): num_dim = input.dim() assert num_dim in [2, 3], f"expected the input to have 2 or 3 dimensions, but got {num_dim} dimensions" l_in = input.shape[-1] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 1 else: return item padding = _convert_int_to_list(self.padding) dilation = _convert_int_to_list(self.dilation) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) l_out = math.floor((l_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) result_shape = tuple(input.shape[:-1]) + (l_out,) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.MaxPool2d) def torch_nn_maxpool2d(self, input): num_dim = input.dim() assert num_dim in [3, 4], f"expected the input to have 3 or 4 dimensions, but got {num_dim} dimensions" h_in, w_in = input.shape[-2:] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 2 else: return item padding = _convert_int_to_list(self.padding) dilation = _convert_int_to_list(self.dilation) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) h_out = math.floor((h_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) w_out = math.floor((w_in + 2 * padding[1] - dilation[1] * (kernel_size[1] - 1) - 1) / stride[1] + 1) result_shape = tuple(input.shape[:-2]) + ( h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.MaxPool3d) def torch_nn_maxpool3d(self, input): num_dim = input.dim() assert num_dim in [4, 5], f"expected the input to have 4 or 5 dimensions, but got {num_dim} dimensions" d_in, h_in, w_in = input.shape[-3:] def _convert_int_to_list(item): if isinstance(item, int): return [item] * 3 else: return item padding = _convert_int_to_list(self.padding) dilation = _convert_int_to_list(self.dilation) kernel_size = _convert_int_to_list(self.kernel_size) stride = _convert_int_to_list(self.stride) d_out = math.floor((d_in + 2 * padding[0] - dilation[0] * (kernel_size[0] - 1) - 1) / stride[0] + 1) h_out = math.floor((h_in + 2 * padding[1] - dilation[1] * (kernel_size[1] - 1) - 1) / stride[1] + 1) w_out = math.floor((w_in + 2 * padding[2] - dilation[2] * (kernel_size[2] - 1) - 1) / stride[2] + 1) result_shape = tuple(input.shape[:-3]) + ( d_out, h_out, w_out, ) return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.AdaptiveAvgPool1d) @meta_patched_module.register(torch.nn.AdaptiveMaxPool1d) def torch_nn_adapative_pooling_1d(self, input): assert input.dim() in [2, 3] if isinstance(self.output_size, int): output_size = (self.output_size,) else: output_size = self.output_size result_shape = tuple(input.shape[:-1]) + output_size return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.AdaptiveAvgPool2d) @meta_patched_module.register(torch.nn.AdaptiveMaxPool2d) def torch_nn_adapative_pooling_2d(self, input): assert input.dim() in [3, 4] if isinstance(self.output_size, int): output_size = (self.output_size,) * 2 else: output_size = self.output_size result_shape = tuple(input.shape[:-2]) + output_size return torch.empty(result_shape, device="meta") @meta_patched_module.register(torch.nn.AdaptiveAvgPool3d) @meta_patched_module.register(torch.nn.AdaptiveMaxPool3d) def torch_nn_adapative_pooling_3d(self, input): assert input.dim() in [4, 5] if isinstance(self.output_size, int): output_size = (self.output_size,) * 3 else: output_size = self.output_size result_shape = tuple(input.shape[:-3]) + output_size return torch.empty(result_shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/rnn.py
colossalai/fx/tracer/meta_patch/patched_module/rnn.py
import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.GRU) @meta_patched_module.register(torch.nn.RNN) def torch_nn_rnn(self, input, hx): assert ( input.shape[-1] == self.input_size ), f"Expected input to have input size {self.input_size} but got {input.shape[-1]} for the torch.nn.RNN patch" assert ( hx.shape[-1] == self.hidden_size ), f"Expected hx to have hidden size {self.hidden_size} but got {hx.shape[-1]} for the torch.nn.RNN patch" d = 2 if self.bidirectional else 1 return torch.empty(input.shape[:-1] + (self.hidden_size * d,), device="meta"), hx
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/activation_function.py
colossalai/fx/tracer/meta_patch/patched_module/activation_function.py
import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.ReLU) @meta_patched_module.register(torch.nn.Sigmoid) @meta_patched_module.register(torch.nn.GELU) @meta_patched_module.register(torch.nn.Tanh) @meta_patched_module.register(torch.nn.ReLU6) @meta_patched_module.register(torch.nn.PReLU) def torch_nn_non_linear_act(self, input): return torch.empty(input.shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/__init__.py
colossalai/fx/tracer/meta_patch/patched_module/__init__.py
from .activation_function import * from .convolution import * from .embedding import * from .linear import * from .normalization import * from .pooling import * from .rnn import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/embedding.py
colossalai/fx/tracer/meta_patch/patched_module/embedding.py
import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.Embedding) def torch_nn_embedding(self, input): result_shape = input.shape + (self.embedding_dim,) return torch.empty(result_shape, device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/normalization.py
colossalai/fx/tracer/meta_patch/patched_module/normalization.py
import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.LayerNorm) @meta_patched_module.register(torch.nn.GroupNorm) @meta_patched_module.register(torch.nn.BatchNorm1d) @meta_patched_module.register(torch.nn.BatchNorm2d) @meta_patched_module.register(torch.nn.BatchNorm3d) def torch_nn_normalize(self, input): # check shape if isinstance(self, torch.nn.BatchNorm1d): assert input.dim() in [2, 3] elif isinstance(self, torch.nn.BatchNorm2d): assert input.dim() == 4 elif isinstance(self, torch.nn.BatchNorm3d): assert input.dim() == 5 # normalization maintain the same shape as the input return input.clone() try: import apex meta_patched_module.register(apex.normalization.FusedLayerNorm)(torch_nn_normalize) meta_patched_module.register(apex.normalization.FusedRMSNorm)(torch_nn_normalize) meta_patched_module.register(apex.normalization.MixedFusedLayerNorm)(torch_nn_normalize) meta_patched_module.register(apex.normalization.MixedFusedRMSNorm)(torch_nn_normalize) except (ImportError, AttributeError): pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/meta_patch/patched_module/linear.py
colossalai/fx/tracer/meta_patch/patched_module/linear.py
import torch from ...registry import meta_patched_module @meta_patched_module.register(torch.nn.Linear) def torch_nn_linear(self, input): last_dim = input.shape[-1] assert ( last_dim == self.in_features ), f"Expected hidden size {self.in_features} but got {last_dim} for the torch.nn.Linear patch" return torch.empty(input.shape[:-1] + (self.out_features,), device="meta")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/__init__.py
colossalai/fx/tracer/bias_addition_patch/__init__.py
from .patched_bias_addition_function import * from .patched_bias_addition_module import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/conv.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/conv.py
import torch from torch.nn.modules.utils import _pair, _single, _triple from ...registry import bias_addition_module from .bias_addition_module import BiasAdditionModule @bias_addition_module.register(torch.nn.Conv1d) @bias_addition_module.register(torch.nn.Conv2d) @bias_addition_module.register(torch.nn.Conv3d) class BiasAdditionConv(BiasAdditionModule): def extract_kwargs_from_mod(self): root = self.tracer.root conv_module = root.get_submodule(self.target) kwarg_attributes = ["groups", "dilation", "stride"] non_bias_kwargs = {} for attr_name in kwarg_attributes: if hasattr(conv_module, attr_name): non_bias_kwargs[attr_name] = getattr(conv_module, attr_name) if conv_module.padding_mode != "zeros": # TODO: non zeros mode requires some extra processing for input conv_type = type(conv_module) if conv_type == "torch.nn.Conv1d": padding_element = _single(0) elif conv_type == "torch.nn.Conv2d": padding_element = _pair(0) elif conv_type == "torch.nn.Conv3d": padding_element = _triple(0) non_bias_kwargs["padding"] = padding_element else: non_bias_kwargs["padding"] = getattr(conv_module, "padding") return non_bias_kwargs def create_bias_reshape_proxy(self, dimensions): """ This method is used to reshape the bias node in order to make bias and output of non-bias convolution broadcastable. """ bias_shape = [1] * (dimensions - 1) bias_shape[0] = -1 bias_reshape_node_kind = "call_method" bias_reshape_node_target = "view" bias_reshape_node_args = (self.bias_proxy, torch.Size(bias_shape)) bias_reshape_proxy = self.tracer.create_proxy( bias_reshape_node_kind, bias_reshape_node_target, bias_reshape_node_args, {} ) return bias_reshape_proxy def generate(self): non_bias_conv_func_proxy = self.create_non_bias_func_proxy() output_dims = non_bias_conv_func_proxy.meta_data.dim() bias_reshape_proxy = self.create_bias_reshape_proxy(output_dims) bias_addition_proxy = self.create_bias_addition_proxy(non_bias_conv_func_proxy, bias_reshape_proxy) return bias_addition_proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/bias_addition_module.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/bias_addition_module.py
import operator from abc import ABC, abstractmethod import torch import torch.nn.functional as F class BiasAdditionModule(ABC): """ This class is used to construct the restructure computation graph for call_module node with bias addition inside. """ def __init__(self, tracer, target, args, kwargs, substitute_func): self.tracer = tracer self.target = target self.args = args self.kwargs = kwargs self.substitute_func = substitute_func self.weight_proxy = self._create_weight_proxy() self.bias_proxy = self._create_bias_proxy() def _create_weight_proxy(self): """ Create weight proxy, the node created by this proxy contains module weight. Note: this function will be invoked during module initializing, you should never call this function. """ weight_node_kind = "get_attr" weight_node_target = self.target + ".weight" weight_proxy = self.tracer.create_proxy(weight_node_kind, weight_node_target, (), {}) return weight_proxy def _create_bias_proxy(self): """ Create bias proxy, the node created by this proxy contains module bias. Note: this function will be invoked during module initializing, you should never call this function. """ bias_node_kind = "get_attr" bias_node_target = self.target + ".bias" bias_proxy = self.tracer.create_proxy(bias_node_kind, bias_node_target, (), {}) return bias_proxy @abstractmethod def extract_kwargs_from_mod(self): """ This method is used to extract the kwargs for non-bias computation. For example: The kwargs for conv2d module is {} because the attributes like 'padding' or 'groups' are considered during module initializing. However, we need to consider those attributes as kwargs in F.conv2d. """ def create_non_bias_func_proxy(self, input_proxy=None): """ This method is used to create the non_bias_func proxy, the node created by this proxy will compute the main computation, such as convolution, with bias option banned. """ node_kind = "call_function" node_target = self.substitute_func if input_proxy is None: input_proxy = self.args[0] node_args = (input_proxy, self.weight_proxy) node_kwargs = self.extract_kwargs_from_mod() non_bias_func_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return non_bias_func_proxy def create_bias_addition_proxy(self, non_bias_func_proxy, bias_proxy): """ This method is used to create the bias_addition_proxy, the node created by this proxy will compute the sum of non_bias_func result and bias with some reshape operation if needed. """ bias_add_node_kind = "call_function" bias_add_node_target = operator.add bias_add_args = (non_bias_func_proxy, bias_proxy) bias_add_proxy = self.tracer.create_proxy(bias_add_node_kind, bias_add_node_target, tuple(bias_add_args), {}) return bias_add_proxy @abstractmethod def generate(self): """ This method is used to construct the whole restructure computation graph for call_module node with bias addition inside. A whole restructure computation graph will contain a weight node, a bias node, a non-bias addition computation node, a bias reshape node if needed and a bias addition node. Use Conv2d module as an example: The origin node is: %conv: call_module[target=conv](args = (%x,), kwargs = {}) Restructured graph is: %conv_weight : [#users=1] = get_attr[target=conv.weight] %conv_bias : [#users=1] = get_attr[target=conv.bias] %conv2d : [#users=1] = call_function[target=torch.conv2d](args = (%x, %conv_weight), kwargs = {}) %view : [#users=1] = call_method[target=view](args = (%conv_bias, [1, -1, 1, 1]), kwargs = {}) %add : [#users=1] = call_function[target=operator.add](args = (%conv2d, %view), kwargs = {}) """ module_to_func_dict = { torch.nn.Linear: F.linear, torch.nn.Conv1d: F.conv1d, torch.nn.Conv2d: F.conv2d, torch.nn.Conv3d: F.conv3d, }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/__init__.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/__init__.py
from .bias_addition_module import * from .conv import * from .linear import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/linear.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_module/linear.py
import torch from ...registry import bias_addition_module from .bias_addition_module import BiasAdditionModule @bias_addition_module.register(torch.nn.Linear) class BiasAdditionLinear(BiasAdditionModule): def extract_kwargs_from_mod(self): return {} def generate(self): non_bias_linear_func_proxy = self.create_non_bias_func_proxy() bias_addition_proxy = self.create_bias_addition_proxy(non_bias_linear_func_proxy, self.bias_proxy) return bias_addition_proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/addbmm.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/addbmm.py
import torch from ...registry import bias_addition_function, bias_addition_method from .bias_addition_function import LinearBasedBiasFunc @bias_addition_method.register(torch.Tensor.addbmm) @bias_addition_function.register(torch.addbmm) class Addbmm(LinearBasedBiasFunc): def extract_kwargs_from_origin_func(self): kwargs = {} if "beta" in self.kwargs: kwargs["beta"] = self.kwargs["beta"] if "alpha" in self.kwargs: kwargs["alpha"] = self.kwargs["alpha"] return kwargs def create_non_bias_func_proxy(self, input_proxy, other_proxy): """ This method is used to create the non_bias_func proxy, the node created by this proxy will compute the main computation, such as convolution, with bias option banned. """ assert self.substitute_func == torch.bmm node_kind = "call_function" node_target = self.substitute_func node_args = (input_proxy, other_proxy) # torch.bmm does not have any kwargs node_kwargs = {} non_bias_func_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return non_bias_func_proxy def insert_sum_node(self, input_proxy, sum_dims=0): """ This method is used to sum the input_proxy through the sum_dims. """ node_kind = "call_function" node_target = torch.sum node_args = (input_proxy, sum_dims) node_kwargs = {} sum_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return sum_proxy def generate(self): # The formula for addbmm is output = beta * input + alpha * (torch.bmm(b1, b2)) # doing the non-bias computation(temp_0 = torch.bmm(b1, b2)) non_bias_linear_func_proxy = self.create_non_bias_func_proxy(self.args[1], self.args[2]) # doing sum on the batch dimension(temp_1 = torch.sum(temp_0, 0)) sum_proxy = self.insert_sum_node(non_bias_linear_func_proxy) kwargs = self.extract_kwargs_from_origin_func() if "beta" in kwargs: beta = kwargs["beta"] # doing the multiplication with beta if it exists(temp_2 = beta * input) beta_proxy = self.create_mul_node(self.args[0], beta) else: beta_proxy = self.args[0] if "alpha" in kwargs: alpha = kwargs["alpha"] # doing the multiplication with alpha if it exists(temp_3 = alpha * temp_1) alpha_proxy = self.create_mul_node(alpha, sum_proxy) else: alpha_proxy = sum_proxy # doing the addition(temp_4 = temp_2 + temp_3) bias_addition_proxy = self.create_bias_addition_proxy(alpha_proxy, beta_proxy) return bias_addition_proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/addmm.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/addmm.py
import torch from ...registry import bias_addition_function, bias_addition_method from .bias_addition_function import LinearBasedBiasFunc @bias_addition_method.register(torch.Tensor.addmm) @bias_addition_function.register(torch.addmm) class Addmm(LinearBasedBiasFunc): def extract_kwargs_from_origin_func(self): kwargs = {} if "beta" in self.kwargs: kwargs["beta"] = self.kwargs["beta"] if "alpha" in self.kwargs: kwargs["alpha"] = self.kwargs["alpha"] return kwargs def transpose_other_operand_for_linear(self, other_proxy): """ This method is used to transpose the other operand for linear function. For example: input = torch.rand(3, 4) m1 = torch.rand(3, 5) m2 = torch.rand(5, 4) original_output = torch.addmm(input, m1, m2) # To keep the computation graph consistent with the origin computation graph, we need to transpose the m2 # before we call the linear function. new_output = torch.linear(m1, m2.transpose(0, 1)) + input """ node_kind = "call_function" node_target = torch.transpose node_args = (other_proxy, 0, 1) node_kwargs = {} transpose_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return transpose_proxy def generate(self): transpose_proxy = self.transpose_other_operand_for_linear(self.args[2]) non_bias_linear_func_proxy = self.create_non_bias_func_proxy(self.args[1], transpose_proxy) kwargs = self.extract_kwargs_from_origin_func() if "beta" in kwargs: beta = kwargs["beta"] beta_proxy = self.create_mul_node(self.args[0], beta) else: beta_proxy = self.args[0] if "alpha" in kwargs: alpha = kwargs["alpha"] alpha_proxy = self.create_mul_node(alpha, non_bias_linear_func_proxy) else: alpha_proxy = non_bias_linear_func_proxy bias_addition_proxy = self.create_bias_addition_proxy(alpha_proxy, beta_proxy) return bias_addition_proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/bias_addition_function.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/bias_addition_function.py
import operator from abc import ABC, abstractmethod import torch import torch.nn.functional as F class BiasAdditionFunc(ABC): """ This class is used to construct the restructure computation graph for call_func node with bias addition inside. """ def __init__(self, tracer, target, args, kwargs, substitute_func): self.tracer = tracer self.target = target self.args = args self.kwargs = kwargs self.substitute_func = substitute_func @abstractmethod def extract_kwargs_from_origin_func(self): """ This method is used to extract the kwargs for further graph transform. For example: The formula for torch.addmm is out = beta * input + alpha * (m1 @ m2) The kwargs for addmm function is {beta=1, alpha=1, output=None}, then we need to insert two more operator.mul nodes for the computation graph to compute the final result. """ @abstractmethod def generate(self): """ This method is used to construct the whole restructure computation graph for call_func node with bias addition inside. A whole restructure computation graph will contain a weight node, a bias node, a non-bias addition computation node, a bias reshape node if needed and a bias addition node. Use torch.addmm as an example: The origin node is: %addmm: call_func[target=torch.addmm](args = (%input_1, m1, m2), kwargs = {beta=1, alpha=1}) Restructured graph is: %transpose : [#users=1] = call_function[target=torch.transpose](args = (%m2, 0, 1), kwargs = {}) %linear : [#users=1] = call_function[target=torch._C._nn.linear](args = (%m1, %transpose), kwargs = {}) %mul : [#users=1] = call_function[target=operator.mul](args = (%input_1, 3), kwargs = {}) %mul_1 : [#users=1] = call_function[target=operator.mul](args = (2, %linear), kwargs = {}) %add : [#users=1] = call_function[target=operator.add](args = (%mul_1, %mul), kwargs = {}) """ def create_mul_node(self, input_proxy, coefficent): """ This method is used to create a coefficent node for the numerical correctness. The formula for torch.addmm is out = beta * input + alpha * (m1 @ m2) Therefore, we need to use this method insert two more operator.mul nodes for the computation graph to compute the final result. """ node_kind = "call_function" node_target = operator.mul node_args = ( input_proxy, coefficent, ) node_kwargs = {} mul_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return mul_proxy class LinearBasedBiasFunc(BiasAdditionFunc): """ This class is used to construct the restructure computation graph for call_func node based on F.linear. """ def create_non_bias_func_proxy(self, input_proxy, other_proxy): """ This method is used to create the non_bias_func proxy, the node created by this proxy will compute the main computation, such as convolution, with bias option banned. """ assert self.substitute_func == torch.nn.functional.linear node_kind = "call_function" node_target = self.substitute_func node_args = (input_proxy, other_proxy) # non-bias linear does not have any kwargs node_kwargs = {} non_bias_func_proxy = self.tracer.create_proxy(node_kind, node_target, node_args, node_kwargs) return non_bias_func_proxy def create_bias_addition_proxy(self, non_bias_func_proxy, bias_proxy): """ This method is used to create the bias_addition_proxy, the node created by this proxy will compute the sum of non_bias_func result and bias with some reshape operation if needed. """ bias_add_node_kind = "call_function" bias_add_node_target = operator.add bias_add_args = (non_bias_func_proxy, bias_proxy) bias_add_proxy = self.tracer.create_proxy(bias_add_node_kind, bias_add_node_target, tuple(bias_add_args), {}) return bias_add_proxy func_to_func_dict = { torch.addmm: F.linear, torch.addbmm: torch.bmm, F.linear: F.linear, } method_to_func_dict = { torch.Tensor.addmm: F.linear, torch.Tensor.addbmm: torch.bmm, }
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/__init__.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/__init__.py
from .addbmm import Addbmm from .addmm import Addmm from .bias_addition_function import BiasAdditionFunc, LinearBasedBiasFunc, func_to_func_dict, method_to_func_dict from .linear import Linear
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/linear.py
colossalai/fx/tracer/bias_addition_patch/patched_bias_addition_function/linear.py
import torch.nn.functional as F from ...registry import bias_addition_function from .bias_addition_function import LinearBasedBiasFunc @bias_addition_function.register(F.linear) class Linear(LinearBasedBiasFunc): def extract_kwargs_from_origin_func(self): assert "bias" in self.kwargs kwargs = {} if "bias" in self.kwargs: kwargs["bias"] = self.kwargs["bias"] return kwargs def generate(self): non_bias_linear_func_proxy = self.create_non_bias_func_proxy(self.args[0], self.args[1]) kwargs = self.extract_kwargs_from_origin_func() bias_addition_proxy = self.create_bias_addition_proxy(non_bias_linear_func_proxy, kwargs["bias"]) return bias_addition_proxy
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/codegen/activation_checkpoint_codegen.py
colossalai/fx/codegen/activation_checkpoint_codegen.py
from typing import Any, Dict, Iterable, List, Tuple import torch import colossalai try: from torch.fx.graph import ( CodeGen, PythonCode, _custom_builtins, _CustomBuiltin, _format_target, _is_from_torch, _Namespace, _origin_type_map, inplace_methods, magic_methods, ) from torch.fx.node import Argument, Node, _get_qualified_name, _type_repr, map_arg CODEGEN_AVAILABLE = True except: from torch.fx.graph import ( PythonCode, _custom_builtins, _CustomBuiltin, _format_args, _format_target, _is_from_torch, _Namespace, _origin_type_map, magic_methods, ) from torch.fx.node import Argument, Node, _get_qualified_name, _type_repr, map_arg CODEGEN_AVAILABLE = False if CODEGEN_AVAILABLE: __all__ = ["ActivationCheckpointCodeGen"] else: __all__ = ["python_code_with_activation_checkpoint"] def _gen_saved_tensors_hooks(): """ Generate saved tensors hooks """ pack_hook = """def pack_hook_input(self, x): if getattr(x, "offload", False): return (x.device, x.cpu()) else: return x def pack_hook_no_input(self, x): if getattr(x, "offload", True): return (x.device, x.cpu()) else: return x """ unpack_hook = """def unpack_hook(self, packed): if isinstance(packed, tuple): device, tensor = packed return tensor.to(device) else: return packed """ return pack_hook, unpack_hook def _gen_save_tensors_hooks_context(offload_input=True) -> str: """Generate customized saved_tensors_hooks Args: offload_input (bool, optional): whether we need offload input, if offload_input=False, we will use self.pack_hook_no_input instead. Defaults to True. Returns: str: generated context """ if offload_input: context = "with torch.autograd.graph.saved_tensors_hooks(self.pack_hook_input, self.unpack_hook):\n" else: context = "with torch.autograd.graph.saved_tensors_hooks(self.pack_hook_no_input, self.unpack_hook):\n" return context def _gen_save_on_cpu_context(): """ Generate save on cpu context """ context = "with torch.autograd.graph.save_on_cpu(pin_memory=True):\n" return context def _find_input_and_output_nodes(nodes: List[Node]): """ Find the input and output node names which are not found in the given list of nodes. """ input_nodes = [] output_nodes = [] # if a node has an input node which is not in the node list # we treat that input node as the input of the checkpoint function for node in nodes: for input_node in node._input_nodes.keys(): node_repr = repr(input_node) if input_node not in nodes and node_repr not in input_nodes: input_nodes.append(node_repr) # if a node has a user node which is not in the node list # we treat that user node as the node receiving the current node output for node in nodes: for output_node in node.users.keys(): node_repr = repr(node) if output_node not in nodes and node_repr not in output_nodes: output_nodes.append(node_repr) return input_nodes, output_nodes def _find_ckpt_regions(nodes: List[Node]): """ Find the checkpoint regions given a list of consecutive nodes. The outputs will be list of tuples, each tuple is in the form of (start_index, end_index). """ ckpt_regions = [] start = -1 end = -1 current_region = None for idx, node in enumerate(nodes): if "activation_checkpoint" in node.meta: act_ckpt_label = node.meta["activation_checkpoint"] # this activation checkpoint label is not set yet # meaning this is the first node of the activation ckpt region if current_region is None: current_region = act_ckpt_label start = idx # if activation checkpoint has changed # we restart the tracking # e.g. node ckpt states = [ckpt1, ckpt2, ckpt2, ckpt2] if act_ckpt_label != current_region: assert start != -1 ckpt_regions.append((start, idx - 1)) current_region = act_ckpt_label start = idx end = -1 elif current_region is not None and not "activation_checkpoint" in node.meta: # used to check the case below # node ckpt states = [ckpt, ckpt, non-ckpt] end = idx - 1 assert start != -1 and end != -1 ckpt_regions.append((start, end)) start = end = -1 current_region = None else: pass return ckpt_regions def _find_offload_regions(nodes: List[Node]): """This function is to find the offload regions In pofo algorithm, during annotation, we will annotate the offload region with the list in the form of [idx, offload_input, offload_bar]. idx indicates the offload region's index, offload_input is a bool type indicates whether we need to offload the input, offload_bar is a bool type indicates whether we need to offload all the intermediate x_bars of this region. """ offload_regions = [] offload_labels = [] start = -1 end = -1 current_region = None for idx, node in enumerate(nodes): if "activation_offload" in node.meta and isinstance(node.meta["activation_offload"], Iterable): act_offload_label = node.meta["activation_offload"] if current_region == None: current_region = act_offload_label start = idx offload_labels.append(act_offload_label) if act_offload_label != current_region: assert start != -1 offload_regions.append((start, idx - 1)) offload_labels.append(act_offload_label) current_region = act_offload_label start = idx end = -1 else: if current_region is not None: end = idx - 1 assert start != -1 and end != -1 offload_regions.append((start, end)) start = end = -1 current_region = None else: pass return offload_regions, offload_labels def _gen_ckpt_fn_def(label, free_vars: List[str]) -> str: """ Generate the checkpoint function definition """ return f"def checkpoint_{label}({', '.join(['self'] + free_vars)}):" def _gen_ckpt_output(output_vars: List[str]) -> str: """ Generate the return statement for checkpoint region """ return f"return {', '.join(output_vars)}" def _gen_ckpt_usage(label, activation_offload, input_vars, output_vars, use_reentrant=True): """ Generate the checkpoint function call code text """ outputs = ", ".join(output_vars) inputs = ", ".join(input_vars) return f"{outputs} = colossalai.utils.activation_checkpoint.checkpoint(self.checkpoint_{label}, {activation_offload}, {inputs}, use_reentrant={use_reentrant})" def _end_of_ckpt(node: Node, check_idx: int) -> bool: """Check if the node could end the ckpt region Args: node (Node): torch.fx.Node check_idx (int): the index of checkpoint level for nested checkpoint Returns: bool """ if "activation_checkpoint" in node.meta: if isinstance(node.meta["activation_checkpoint"], list): return node.meta["activation_checkpoint"][check_idx] == None else: return False else: return True def _find_nested_ckpt_regions(nodes, check_idx=0): """ Find the nested checkpoint regions given a list of consecutive nodes. The outputs will be list of tuples, each tuple is in the form of (start_index, end_index). """ ckpt_regions = [] start = -1 end = -1 current_region = None for idx, node in enumerate(nodes): if "activation_checkpoint" in node.meta: if isinstance(node.meta["activation_checkpoint"], int): act_ckpt_label = node.meta["activation_checkpoint"] else: act_ckpt_label = node.meta["activation_checkpoint"][check_idx] # this activation checkpoint label is not set yet # meaning this is the first node of the activation ckpt region if current_region is None: current_region = act_ckpt_label start = idx # if activation checkpoint has changed # we restart the tracking # e.g. node ckpt states = [ckpt1, ckpt2, ckpt2, ckpt2] if act_ckpt_label != current_region: assert start != -1 ckpt_regions.append((start, idx - 1)) current_region = act_ckpt_label start = idx end = -1 elif current_region is not None and _end_of_ckpt(node, check_idx): # used to check the case below # node ckpt states = [ckpt, ckpt, non-ckpt] end = idx - 1 assert start != -1 and end != -1 ckpt_regions.append((start, end)) start = end = -1 current_region = None else: pass if current_region is not None: end = len(nodes) - 1 ckpt_regions.append((start, end)) return ckpt_regions def emit_ckpt_func( body, ckpt_func, node_list: List[Node], emit_node_func, delete_unused_value_func, level=0, in_ckpt=False ): """Emit ckpt function in nested way Args: body: forward code, in recursive calls, this part will be checkpoint functions code ckpt_func: checkpoint functions code, in recursive calls, this part will be a buffer node_list (List[Node]): list of torch.fx.Node emit_node_func: function to emit a node delete_unused_value_func: function to delete unused value level (int, optional): checkpoint level. Defaults to 0. in_ckpt (bool, optional): indicates wether the func is in recursive call. Defaults to False. """ inputs, outputs = _find_input_and_output_nodes(node_list) # if the current checkpoint function use int as label, using old generation method if isinstance(node_list[0].meta["activation_checkpoint"], int): label = node_list[0].meta["activation_checkpoint"] ckpt_fn_def = _gen_ckpt_fn_def(label, inputs) ckpt_func.append(f"{ckpt_fn_def}\n") for node in node_list: emit_node_func(node, ckpt_func) ckpt_func[-1] = " " + ckpt_func[-1] delete_unused_value_func(node, ckpt_func) ckpt_func.append(" " + _gen_ckpt_output(outputs) + "\n\n") activation_offload = node_list[0].meta.get("activation_offload", False) usage = _gen_ckpt_usage(label, activation_offload, inputs, outputs, False) usage += "\n" body.append(usage) # use nested ckpt function codegen else: # label given by each layer, e.g. if you are currently at level [0, 1, 1] # the label will be '0_1_1' label = "_".join([str(idx) for idx in node_list[0].meta["activation_checkpoint"][: level + 1]]) ckpt_fn_def = _gen_ckpt_fn_def(label, inputs) ckpt_func.append(f"{ckpt_fn_def}\n") # if there is more level to fetch if level + 1 < len(node_list[0].meta["activation_checkpoint"]): ckpt_regions = _find_nested_ckpt_regions(node_list, level + 1) start_idx = [item[0] for item in ckpt_regions] end_idx = [item[1] for item in ckpt_regions] # use ckpt_func_buffer to store nested checkpoint functions ckpt_func_buffer = [] node_idx = 0 while 1: if node_idx >= len(node_list): break if node_idx in start_idx: ckpt_node_list = node_list[node_idx : end_idx[start_idx.index(node_idx)] + 1] emit_ckpt_func( ckpt_func, ckpt_func_buffer, ckpt_node_list, emit_node_func, delete_unused_value_func, level + 1, True, ) node_idx += len(ckpt_node_list) else: node = node_list[node_idx] emit_node_func(node, ckpt_func) ckpt_func[-1] = " " + ckpt_func[-1] delete_unused_value_func(node, ckpt_func) node_idx += 1 ckpt_func.append(" " + _gen_ckpt_output(outputs) + "\n\n") ckpt_func += ckpt_func_buffer activation_offload = node_list[0].meta.get("activation_offload", False) usage = _gen_ckpt_usage(label, activation_offload, inputs, outputs, False) + "\n" if in_ckpt: usage = " " + usage body.append(usage) # last level else: for node in node_list: emit_node_func(node, ckpt_func) ckpt_func[-1] = " " + ckpt_func[-1] delete_unused_value_func(node, ckpt_func) ckpt_func.append(" " + _gen_ckpt_output(outputs) + "\n\n") activation_offload = node_list[0].meta.get("activation_offload", False) usage = _gen_ckpt_usage(label, activation_offload, inputs, outputs, False) + "\n" if in_ckpt: usage = " " + usage body.append(usage) def emit_code_with_nested_activation_checkpoint(body, ckpt_func, nodes, emit_node_func, delete_unused_value_func): """Emit code with nested activation checkpoint When we detect some of the node.activation_checkpoint is a List, we will use this function to emit the activation checkpoint codes. Args: body: forward code ckpt_func: checkpoint functions code nodes: graph.nodes emit_node_func: function to emit node delete_unused_value_func: function to remove the unused value """ ckpt_regions = _find_nested_ckpt_regions(nodes, 0) start_idx = [item[0] for item in ckpt_regions] end_idx = [item[1] for item in ckpt_regions] # find the offload regions offload_regions, offload_labels = _find_offload_regions(nodes) offload_starts = [item[0] for item in offload_regions] offload_ends = [item[1] for item in offload_regions] offload_inputs = [] offload_outputs = [] within_offload_region = False node_list = list(nodes) # find the input and output var names for each offload region for idx, (start, end) in enumerate(offload_regions): offload_node_list = node_list[start : end + 1] inputs, outputs = _find_input_and_output_nodes(offload_node_list) offload_inputs.append(inputs) offload_outputs.append(outputs) # this flag is to prevent repeated insert of save tensors # hooks definition in ckpt_func is_hook_inserted = False node_idx = 0 while 1: # break if we finish the processing all the nodes if node_idx >= len(node_list): break # process ckpt_regions if node_idx in start_idx: ckpt_node_list = node_list[node_idx : end_idx[start_idx.index(node_idx)] + 1] emit_ckpt_func(body, ckpt_func, ckpt_node_list, emit_node_func, delete_unused_value_func) node_idx += len(ckpt_node_list) # process node in forward function else: node = node_list[node_idx] if node_idx in offload_starts: offload_label = offload_labels[offload_starts.index(node_idx)] _, offload_input, offload_bar = offload_label within_offload_region = True # insert hook functions if needed if not is_hook_inserted: pack_hook, unpack_hook = _gen_saved_tensors_hooks() ckpt_func.insert(0, "\n".join([pack_hook, unpack_hook]) + "\n") is_hook_inserted = True if offload_input and offload_bar: body.append(_gen_save_on_cpu_context()) elif offload_input: for par in offload_inputs[offload_label[0]]: body.append(f"setattr({par}, 'offload', True)\n") body.append(_gen_save_tensors_hooks_context(offload_input=True)) else: for par in offload_inputs[offload_label[0]]: body.append(f"setattr({par}, 'offload', False)\n") body.append(_gen_save_tensors_hooks_context(offload_input=False)) if within_offload_region: emit_node_func(node, body) body[-1] = " " + body[-1] delete_unused_value_func(node, body) else: emit_node_func(node, body) delete_unused_value_func(node, body) if node_idx in offload_ends: within_offload_region = False node_idx += 1 def emit_code_with_activation_checkpoint(body, ckpt_func, nodes, emit_node_func, delete_unused_value_func): # find the activation checkpoint regions ckpt_regions = _find_ckpt_regions(nodes) start_idx = [item[0] for item in ckpt_regions] end_idx = [item[1] for item in ckpt_regions] input_vars = [] output_vars = [] within_ckpt_region = False # find the offload regions offload_regions, offload_labels = _find_offload_regions(nodes) offload_starts = [item[0] for item in offload_regions] offload_ends = [item[1] for item in offload_regions] offload_inputs = [] offload_outputs = [] within_offload_region = False node_list = list(nodes) # use this variable to avoid inserting hook functions # to ckpt_func repeatedly is_hook_inserted = False # find the input and output var names for each region for idx, (start, end) in enumerate(ckpt_regions): ckpt_node_list = node_list[start : end + 1] inputs, outputs = _find_input_and_output_nodes(ckpt_node_list) input_vars.append(inputs) output_vars.append(outputs) # find the input and output var names for each offload region for idx, (start, end) in enumerate(offload_regions): offload_node_list = node_list[start : end + 1] inputs, outputs = _find_input_and_output_nodes(offload_node_list) offload_inputs.append(inputs) offload_outputs.append(outputs) # append code text to body for idx, node in enumerate(node_list): # if this is the first node of the ckpt region # append the ckpt function definition if idx in start_idx: label = start_idx.index(idx) ckpt_fn_def = _gen_ckpt_fn_def(label, input_vars[label]) ckpt_func.append(f"{ckpt_fn_def}\n") within_ckpt_region = True if idx in offload_starts: offload_label = offload_labels[offload_starts.index(idx)] _, offload_input, offload_bar = offload_label within_offload_region = True # insert hook functions if needed if not is_hook_inserted: pack_hook, unpack_hook = _gen_saved_tensors_hooks() ckpt_func.insert(0, "\n".join([pack_hook, unpack_hook]) + "\n") is_hook_inserted = True if offload_input and offload_bar: body.append(_gen_save_on_cpu_context()) elif offload_input: for par in offload_inputs[offload_label[0]]: body.append(f"setattr({par}, 'offload', True)\n") body.append(_gen_save_tensors_hooks_context(offload_input=True)) else: for par in offload_inputs[offload_label[0]]: body.append(f"setattr({par}, 'offload', False)\n") body.append(_gen_save_tensors_hooks_context(offload_input=False)) # NOTE: emit_node does not emit a string with newline. It depends # on delete_unused_values to append one # NOTE: currently we separate body and ckpt_func definition if within_ckpt_region: emit_node_func(node, ckpt_func) ckpt_func[-1] = " " + ckpt_func[-1] delete_unused_value_func(node, ckpt_func) elif within_offload_region: emit_node_func(node, body) body[-1] = " " + body[-1] delete_unused_value_func(node, body) else: emit_node_func(node, body) delete_unused_value_func(node, body) if idx in end_idx: # if this is the last node of the ckpt region # generate return statement label = end_idx.index(idx) return_statement = _gen_ckpt_output(output_vars[label]) return_statement = f" {return_statement}\n\n" ckpt_func.append(return_statement) # we need to check if the checkpoint need to offload the input start_node_idx = start_idx[label] if "activation_offload" in node_list[start_node_idx].meta: activation_offload = node_list[start_node_idx].meta["activation_offload"] else: activation_offload = False # we need to check if the checkpoint need use_reentrant=False use_reentrant = True non_leaf_input = 0 for var in input_vars[label]: input_node = next(item for item in node_list if item.name == var) if input_node.op != "placeholder": non_leaf_input = 1 for user in input_node.users: if "activation_checkpoint" in user.meta: if user.meta["activation_checkpoint"] == label: if user.op == "call_module": if hasattr(user.graph.owning_module.get_submodule(user.target), "inplace"): use_reentrant = not user.graph.owning_module.get_submodule(user.target).inplace elif user.op == "call_function": if "inplace" in user.kwargs: use_reentrant = not user.kwargs["inplace"] # if all the inputs are leaf nodes, we need to set use_reentrant = False if not non_leaf_input: use_reentrant = False # generate checkpoint function call in a new line usage = _gen_ckpt_usage(label, activation_offload, input_vars[label], output_vars[label], use_reentrant) usage += "\n" body.append(usage) within_ckpt_region = False if idx in offload_ends: within_offload_region = False if CODEGEN_AVAILABLE: class ActivationCheckpointCodeGen(CodeGen): def _gen_python_code(self, nodes, root_module: str, namespace: _Namespace, verbose=None) -> PythonCode: free_vars: List[str] = [] body: List[str] = [] globals_: Dict[str, Any] = {} wrapped_fns: Dict[str, None] = {} # Wrap string in list to pass by reference maybe_return_annotation: List[str] = [""] def add_global(name_hint: str, obj: Any): """Add an obj to be tracked as a global. We call this for names that reference objects external to the Graph, like functions or types. Returns: the global name that should be used to reference 'obj' in generated source. """ if _is_from_torch(obj) and obj != torch.device: # to support registering torch.device # HACK: workaround for how torch custom ops are registered. We # can't import them like normal modules so they must retain their # fully qualified name. return _get_qualified_name(obj) # normalize the name hint to get a proper identifier global_name = namespace.create_name(name_hint, obj) if global_name in globals_: assert globals_[global_name] is obj return global_name globals_[global_name] = obj return global_name # set _custom_builtins here so that we needn't import colossalai in forward _custom_builtins["colossalai"] = _CustomBuiltin("import colossalai", colossalai) # Pre-fill the globals table with registered builtins. for name, (_, obj) in _custom_builtins.items(): add_global(name, obj) def type_repr(o: Any): if o == (): # Empty tuple is used for empty tuple type annotation Tuple[()] return "()" typename = _type_repr(o) if hasattr(o, "__origin__"): # This is a generic type, e.g. typing.List[torch.Tensor] origin_type = _origin_type_map.get(o.__origin__, o.__origin__) origin_typename = add_global(_type_repr(origin_type), origin_type) if hasattr(o, "__args__"): # Assign global names for each of the inner type variables. args = [type_repr(arg) for arg in o.__args__] if len(args) == 0: # Bare type, such as `typing.Tuple` with no subscript # This code-path used in Python < 3.9 return origin_typename return f'{origin_typename}[{",".join(args)}]' else: # Bare type, such as `typing.Tuple` with no subscript # This code-path used in Python 3.9+ return origin_typename # Common case: this is a regular module name like 'foo.bar.baz' return add_global(typename, o) def _format_args(args: Tuple[Argument, ...], kwargs: Dict[str, Argument]) -> str: def _get_repr(arg): # Handle NamedTuples (if it has `_fields`) via add_global. if isinstance(arg, tuple) and hasattr(arg, "_fields"): qualified_name = _get_qualified_name(type(arg)) global_name = add_global(qualified_name, type(arg)) return f"{global_name}{repr(tuple(arg))}" return repr(arg) args_s = ", ".join(_get_repr(a) for a in args) kwargs_s = ", ".join(f"{k} = {_get_repr(v)}" for k, v in kwargs.items()) if args_s and kwargs_s: return f"{args_s}, {kwargs_s}" return args_s or kwargs_s # Run through reverse nodes and record the first instance of a use # of a given node. This represents the *last* use of the node in the # execution order of the program, which we will use to free unused # values node_to_last_use: Dict[Node, Node] = {} user_to_last_uses: Dict[Node, List[Node]] = {} def register_last_uses(n: Node, user: Node): if n not in node_to_last_use: node_to_last_use[n] = user user_to_last_uses.setdefault(user, []).append(n) for node in reversed(nodes): map_arg(node.args, lambda n: register_last_uses(n, node)) map_arg(node.kwargs, lambda n: register_last_uses(n, node)) # NOTE: we add a variable to distinguish body and ckpt_func def delete_unused_values(user: Node, body): """ Delete values after their last use. This ensures that values that are not used in the remainder of the code are freed and the memory usage of the code is optimal. """ if user.op == "placeholder": return if user.op == "output": body.append("\n") return nodes_to_delete = user_to_last_uses.get(user, []) if len(nodes_to_delete): to_delete_str = " = ".join([repr(n) for n in nodes_to_delete] + ["None"]) body.append(f"; {to_delete_str}\n") else: body.append("\n") # NOTE: we add a variable to distinguish body and ckpt_func def emit_node(node: Node, body): maybe_type_annotation = "" if node.type is None else f" : {type_repr(node.type)}" if node.op == "placeholder": assert isinstance(node.target, str) maybe_default_arg = "" if not node.args else f" = {repr(node.args[0])}" free_vars.append(f"{node.target}{maybe_type_annotation}{maybe_default_arg}") raw_name = node.target.replace("*", "") if raw_name != repr(node): body.append(f"{repr(node)} = {raw_name}\n") return elif node.op == "call_method": assert isinstance(node.target, str) body.append( f"{repr(node)}{maybe_type_annotation} = {_format_target(repr(node.args[0]), node.target)}" f"({_format_args(node.args[1:], node.kwargs)})" ) return elif node.op == "call_function": assert callable(node.target) # pretty print operators if node.target.__module__ == "_operator" and node.target.__name__ in magic_methods: assert isinstance(node.args, tuple) body.append( f"{repr(node)}{maybe_type_annotation} = " f"{magic_methods[node.target.__name__].format(*(repr(a) for a in node.args))}" ) return # pretty print inplace operators; required for jit.script to work properly # not currently supported in normal FX graphs, but generated by torchdynamo if node.target.__module__ == "_operator" and node.target.__name__ in inplace_methods: body.append( f"{inplace_methods[node.target.__name__].format(*(repr(a) for a in node.args))}; " f"{repr(node)}{maybe_type_annotation} = {repr(node.args[0])}" ) return qualified_name = _get_qualified_name(node.target) global_name = add_global(qualified_name, node.target) # special case for getattr: node.args could be 2-argument or 3-argument # 2-argument: attribute access; 3-argument: fall through to attrib function call with default value if ( global_name == "getattr" and isinstance(node.args, tuple) and isinstance(node.args[1], str) and node.args[1].isidentifier() and len(node.args) == 2 ): body.append( f"{repr(node)}{maybe_type_annotation} = {_format_target(repr(node.args[0]), node.args[1])}" ) return body.append( f"{repr(node)}{maybe_type_annotation} = {global_name}({_format_args(node.args, node.kwargs)})" ) if node.meta.get("is_wrapped", False): wrapped_fns.setdefault(global_name) return elif node.op == "call_module": assert isinstance(node.target, str) body.append( f"{repr(node)}{maybe_type_annotation} = " f"{_format_target(root_module, node.target)}({_format_args(node.args, node.kwargs)})" ) return
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/codegen/__init__.py
colossalai/fx/codegen/__init__.py
from .activation_checkpoint_codegen import *
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/shard_1d_pass.py
colossalai/fx/passes/shard_1d_pass.py
import operator import torch import torch.nn as nn from colossalai.legacy.tensor import ProcessGroup from colossalai.legacy.tensor.compute_spec import ComputePattern, ComputeSpec from colossalai.legacy.tensor.distspec import ShardSpec ELEMENTWISE_MODULE_OP = [torch.nn.Dropout, torch.nn.ReLU] ELEMENTWISE_FUNC_OP = [ torch.add, operator.add, torch.abs, torch.cos, torch.exp, torch.mul, operator.mul, operator.floordiv, operator.truediv, operator.neg, torch.multiply, torch.nn.functional.relu, torch.nn.functional.dropout, ] def weight_split(weight: torch.nn.parameter.Parameter, dim: int, col_normal: bool) -> torch.nn.parameter.Parameter: """weight_split split a nn.Parameter Args: weight (torch.nn.parameter.Parameter): a torch Parameter instance dim (int): the dimension to be sharded along with col_normal(bool): col shard with gather or not Returns: _type_: _description_ """ if col_normal: setattr(weight, "fx_attr", (dim, "SHARD", "TP", "col_normal")) else: setattr(weight, "fx_attr", (dim, "SHARD", "TP", "col_needs_many_outputs")) return weight def column_shard_linear_pass(gm: torch.fx.GraphModule): # Split all the linear module with column shard. Currently for testing only. mod_graph = gm.graph for node in mod_graph.nodes: if node.op == "call_module": target_module = node.graph.owning_module.get_submodule(node.target) if isinstance(target_module, torch.nn.Linear): target_module.weight = weight_split(target_module.weight, dim=0, col_normal=False) if target_module.bias is not None: target_module.bias.data = weight_split(target_module.bias.data, dim=0, col_normal=False) gm.recompile() return gm def row_shard_linear_pass(gm: torch.fx.GraphModule): # Split all the linear module with row shard. Currently for testing only. mod_graph = gm.graph for node in mod_graph.nodes: if node.op == "call_module": target_module = node.graph.owning_module.get_submodule(node.target) if isinstance(target_module, torch.nn.Linear): target_module.weight = weight_split(target_module.weight, dim=-1, col_normal=False) gm.recompile() return gm def transformer_mlp_pass(graph_module: torch.fx.GraphModule, process_group: ProcessGroup): """ This IR pass checks for transformer MLP like structure and annotate column and row sharding to the linear layers. """ # TODO: Needs to handle special cases, like x = linear(x) + linear(x) graph = graph_module.graph world_size = process_group.world_size() def _traverse_and_annotate(node, start_tracking, annotation_record, world_size): # traverse the graph to look for consecutive linear layers is_linear_module = False if node.op == "call_module": # look for the linear layer module = node.graph.owning_module.get_submodule(node.target) if isinstance(module, nn.Linear): is_linear_module = True if start_tracking: # when start_tracking = True # it means the first linear has been found and the current module # is the second linear # set the current linear module to be row-sharded annotation_record["row"] = module for shard_type, module in annotation_record.items(): # add row sharding spec if shard_type == "row": dist_spec = ShardSpec(dims=[-1], num_partitions=[world_size]) comp_spec = ComputeSpec(ComputePattern.TP1D) setattr(module.weight, "pg", process_group) setattr(module.weight, "dist_spec", dist_spec) setattr(module.weight, "comp_spec", comp_spec) elif shard_type == "col": weight_dist_spec = ShardSpec(dims=[0], num_partitions=[world_size]) weight_comp_spec = ComputeSpec(ComputePattern.TP1D) weight_comp_spec.output_replicate = False setattr(module.weight, "pg", process_group) setattr(module.weight, "dist_spec", weight_dist_spec) setattr(module.weight, "comp_spec", weight_comp_spec) if module.bias is not None: bias_dist_spec = ShardSpec(dims=[0], num_partitions=[world_size]) bias_comp_spec = ComputeSpec(ComputePattern.TP1D) bias_comp_spec.output_replicate = False setattr(module.bias, "pg", process_group) setattr(module.bias, "dist_spec", bias_dist_spec) setattr(module.bias, "comp_spec", bias_comp_spec) start_tracking = False annotation_record.clear() else: # when start tracking = False # it means the current layer is the first linear # set the linear layer to be col-sharded start_tracking = True annotation_record["col"] = module if start_tracking and not is_linear_module: # check against the white list # if non-element wise op is found, we reset the tracking if node.op == "call_module": module = node.graph.owning_module.get_submodule(node.target) if module.__class__ not in ELEMENTWISE_MODULE_OP: start_tracking = False elif node.op == "call_function" or node.op == "call_method": if node.target not in ELEMENTWISE_FUNC_OP: start_tracking = False elif len(node.users.keys()) > 1: start_tracking = False if not start_tracking: annotation_record.clear() # stop tracking for consecutive linear when branch is found # e.g. # out1 = self.linear1(x) # out2 = self.linear2(x) # return out1+out2 next_nodes = list(node.users.keys()) if len(next_nodes) > 1: start_tracking = False annotation_record.clear() # traverse for node in next_nodes: _traverse_and_annotate(node, start_tracking, annotation_record, world_size) placeholder_node = list(graph.nodes)[0] annotate_record = {} _traverse_and_annotate(placeholder_node, False, annotate_record, world_size) return graph_module
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/adding_split_node_pass.py
colossalai/fx/passes/adding_split_node_pass.py
import numpy as np import torch import tqdm from colossalai.fx.passes.split_module import split_module def pipe_split(): pass def block_split(): pass # Construct blocks with the condition that (block_flops / total_flops) >= limit. def construct_blocks(gm: torch.fx.GraphModule, limit=0.01): total_fwd_flop = 0 total_bwd_flop = 0 for node in gm.graph.nodes: total_fwd_flop += node.fwd_flop total_bwd_flop += node.bwd_flop total_flop = total_fwd_flop + total_bwd_flop per_block_flop = total_flop * limit accumulate_fwd_flop = 0 accumulate_bwd_flop = 0 block_nodes = [] for node in gm.graph.nodes: if "block_split" in node.name: continue accumulate_fwd_flop += node.fwd_flop accumulate_bwd_flop += node.bwd_flop if accumulate_fwd_flop + accumulate_bwd_flop >= per_block_flop: with gm.graph.inserting_after(node): block_node = gm.graph.create_node("call_function", block_split) setattr(block_node, "fwd_flop", accumulate_fwd_flop) setattr(block_node, "bwd_flop", accumulate_bwd_flop) accumulate_fwd_flop = 0 accumulate_bwd_flop = 0 block_nodes.append(block_node) return block_nodes def remove_blocks(gm: torch.fx.GraphModule): for node in gm.graph.nodes: if (node.op, node.target) == ("call_function", block_split): gm.graph.erase_node(node) def get_compute_costs(node_list): num_nodes = len(node_list) all_compute_cost = np.full((num_nodes, num_nodes), np.inf, dtype=np.float64) for start in tqdm.tqdm(range(num_nodes), desc="start pos", position=0): for end in tqdm.tqdm(range(start, num_nodes), desc="end pos", position=1, leave=False): selected_flops = [(node_list[i].fwd_flop + node_list[i].bwd_flop) for i in range(start, end + 1)] all_compute_cost[start, end] = sum(selected_flops) return all_compute_cost def do_dp_split_gpipe_impl(num_nodes, num_stages, num_microbatches, compute_costs, max_compute_cost): """The core implementation of the DP algorithm.""" # Adapted from Alpa DP Formulation. # For f, node ID start from 0 # f[number of stages, # node id that is currently being considered] # record time cost(assess by fwd+bwd flop now) f = np.full((num_stages + 1, num_nodes + 1), np.inf, dtype=np.float32) # record max stage compute cost among all stages in this partition. f_stage_max = np.full((num_stages + 1, num_nodes + 1), 0.0, dtype=np.float32) # record start node index for next stage in this partition f_argmin = np.full((num_stages + 1, num_nodes + 1), -1, dtype=np.int32) f[0, num_nodes] = 0 for s in tqdm.tqdm( range(1, num_stages + 1), desc="stage", position=2, leave=False ): # pylint: disable=too-many-nested-blocks for i in tqdm.tqdm(range(num_nodes - 1, -1, -1), desc="start node", position=3, leave=False): for k in tqdm.tqdm(range(num_nodes, i, -1), desc="mid node", position=4, leave=False): stage_cost = compute_costs[i, k - 1] new_cost = f[s - 1, k] + stage_cost if stage_cost <= max_compute_cost and new_cost < f[s, i]: f[s, i] = new_cost f_stage_max[s, i] = max(f_stage_max[s - 1, k], stage_cost) f_argmin[s, i] = k best_total_cost = f[num_stages, 0] if np.isinf(best_total_cost): return np.inf, None total_cost = f[num_stages, 0] + (num_microbatches - 1) * f_stage_max[num_stages, 0] current_s = num_stages current_node = 0 res = [] while current_s > 0 and current_node < num_nodes: next_start_node = f_argmin[current_s, current_node] res.append((current_node, next_start_node)) current_s -= 1 current_node = next_start_node return total_cost, res def do_dp_split_gpipe(node_list, compute_costs, num_stages: int, num_microbatches: int): # Ignore the memory cost profiling in Alpa's design for convenience. max_compute_costs = np.sort(np.unique(compute_costs)) best_cost = np.inf best_solution = None last_max_compute_cost = 0.0 gap = 1e6 # temporary magic number, unit: flops for max_compute_cost in tqdm.tqdm(max_compute_costs): # Pruning to reduce search space. if max_compute_cost * num_microbatches >= best_cost: break if max_compute_cost - last_max_compute_cost < gap: continue cost, solution = do_dp_split_gpipe_impl( len(node_list), num_stages, num_microbatches, compute_costs, max_compute_cost ) if cost < best_cost: best_cost = cost best_solution = solution last_max_compute_cost = max_compute_cost return best_cost, best_solution # Auto DP partition based on Alpa. # Adapted to Gpipe Scheduler # split_mode: # 'node': fx_node # 'block': many fx_nodes construct a block def gpipe_dp_split_pass(gm: torch.fx.GraphModule, pp_size: int, num_microbatches: int, mode="block", block_limit=0.01): assert mode in ["node", "block"] # nodes or blocks will be used in partition. node_list = [] if mode == "node": for node in gm.graph.nodes: node_list.append(node) elif mode == "block": node_list = construct_blocks(gm, limit=block_limit) else: pass compute_costs = get_compute_costs(node_list) best_cost, best_solution = do_dp_split_gpipe(node_list, compute_costs, pp_size, num_microbatches) for _, next_start_node in best_solution: if pp_size <= 1: break node = node_list[next_start_node] with gm.graph.inserting_before(node): split_node = gm.graph.create_node("call_function", pipe_split) pp_size -= 1 # remove block node if possible if mode == "block": remove_blocks(gm) gm.recompile() return gm def avgcompute_split_pass(gm: torch.fx.GraphModule, pp_size: int): """ In avgcompute_split_pass, we split module by the fwd flops. """ mod_graph = gm.graph # To use avgcompute_split_pass, we need run meta_info_prop interpreter first. # If nodes don't have meta info, this pass will fall back to normal balanced split pass. check_node = list(mod_graph.nodes)[0] if "tensor_meta" not in check_node.meta: return balanced_split_pass(gm, pp_size) total_fwd_flop = 0 for node in mod_graph.nodes: total_fwd_flop += node.fwd_flop partition_flop = total_fwd_flop // pp_size accumulate_fwd_flop = 0 for node in mod_graph.nodes: if pp_size <= 1: break if "pipe_split" in node.name: continue accumulate_fwd_flop += node.fwd_flop if accumulate_fwd_flop >= partition_flop: total_fwd_flop = total_fwd_flop - accumulate_fwd_flop accumulate_fwd_flop = 0 pp_size -= 1 partition_flop = total_fwd_flop // pp_size with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def avgnode_split_pass(gm: torch.fx.GraphModule, pp_size: int): """ In avgnode_split_pass, simply split graph by node number. """ mod_graph = gm.graph avg_num_node = len(mod_graph.nodes) // pp_size accumulate_num_node = 0 for node in mod_graph.nodes: if pp_size <= 1: break accumulate_num_node += 1 if accumulate_num_node >= avg_num_node: accumulate_num_node = 0 pp_size -= 1 if node.next.op == "output": with mod_graph.inserting_before(node): split_node = mod_graph.create_node("call_function", pipe_split) else: with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def balanced_split_pass(gm: torch.fx.GraphModule, pp_size: int): """ In balanced_split_pass, we split module by the size of parameters(weights+bias). """ mod_graph = gm.graph total_param_amount = 0 for param in mod_graph.owning_module.parameters(): total_param_amount += param.numel() params_per_partition = total_param_amount // pp_size accumulate_param_amount = 0 for node in mod_graph.nodes: if pp_size <= 1: break if node.op == "call_module": target_module = node.graph.owning_module.get_submodule(node.target) for param in target_module.parameters(): accumulate_param_amount += param.numel() if accumulate_param_amount >= params_per_partition: accumulate_param_amount = 0 pp_size -= 1 # If the next node is output node, we will insert split annotation before # node to make sure there is at least one node in last partition. if node.next.op == "output": with mod_graph.inserting_before(node): split_node = mod_graph.create_node("call_function", pipe_split) else: with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) if pp_size > 1: node_counter = 0 for node in mod_graph.nodes: if pp_size <= 1: break if node.op == "placeholder": continue elif node_counter == 0: node_counter += 1 else: pp_size -= 1 node_counter = 0 with mod_graph.inserting_before(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def balanced_split_pass_v2(gm: torch.fx.GraphModule, pp_size: int): """ In balanced_split_pass_v12, we split module by the size of nodes(weights+bias+outputs). """ mod_graph = gm.graph # To use balanced_split_pass_v2, we need run meta_info_prop interpreter first. # If nodes don't have meta info, this pass will fall back to normal balanced split pass. check_node = list(mod_graph.nodes)[0] if "tensor_meta" not in check_node.meta: return balanced_split_pass(gm, pp_size) total_element_size = 0 for node in mod_graph.nodes: total_element_size += node.node_size partition_size = total_element_size // pp_size accumulate_node_size = 0 for node in mod_graph.nodes: if pp_size <= 1: break if "pipe_split" in node.name: continue accumulate_node_size += node.node_size if accumulate_node_size >= partition_size: total_element_size = total_element_size - accumulate_node_size accumulate_node_size = 0 pp_size -= 1 partition_size = total_element_size // pp_size with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def uniform_split_pass(gm: torch.fx.GraphModule, pp_size: int): mod_graph = gm.graph valid_children_size = 0 valid_children = [] for module in mod_graph.owning_module.children(): valid_children_size += 1 valid_children.append(module) if valid_children_size < pp_size: # If valid children is not enough to shard, we will use balanced policy instead of uniform policy. return balanced_split_pass(gm, pp_size) layers_per_partition = valid_children_size // pp_size accumulate_layer_amount = 0 for node in mod_graph.nodes: if pp_size <= 1: break if node.op == "call_module": target_module = node.graph.owning_module.get_submodule(node.target) if target_module in valid_children: accumulate_layer_amount += 1 if accumulate_layer_amount == layers_per_partition: accumulate_layer_amount = 0 pp_size -= 1 with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def split_with_split_nodes_pass(annotated_gm: torch.fx.GraphModule, merge_output=False): # TODO(lyl): use partition IR to assign partition ID to each node. # Currently: analyzing graph -> annotate graph by inserting split node -> use split module pass to split graph # In future: graph to partitions -> analyzing partition IR -> recombining partitions to get best performance -> assign partition ID to each node part_idx = 0 def split_callback(n: torch.fx.Node): nonlocal part_idx if (n.op, n.target) == ("call_function", pipe_split): part_idx += 1 return part_idx split_mod = split_module(annotated_gm, None, split_callback, merge_output) split_submodules = [] for name, submodule in split_mod.named_modules(): if isinstance(submodule, torch.fx.GraphModule): for node in submodule.graph.nodes: if (node.op, node.target) == ("call_function", pipe_split): submodule.graph.erase_node(node) submodule.recompile() split_submodules.append(submodule) return split_mod, split_submodules
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/utils.py
colossalai/fx/passes/utils.py
from typing import Dict import torch from torch.fx.graph import Graph from torch.fx.node import Node, map_arg def get_comm_size(prev_partition, next_partition): """ Given two partitions (parent and child), calculate the communication size between the two. """ # Keep tracking the communication size between parent and child comm_size = 0 # Keep tracking all the counted node visited_nodes = set() # Go through all nodes in the child partition # If a node has input nodes from the parent partition, # the output size of those input nodes will be counted # and added to comm_size parent_node_names = [n.name for n in prev_partition.graph.nodes] for node in next_partition.graph.nodes: input_nodes: Dict[Node, None] = {} map_arg(node.args, lambda n: input_nodes.setdefault(n)) map_arg(node.kwargs, lambda n: input_nodes.setdefault(n)) for n in input_nodes: if n.name in parent_node_names and n not in visited_nodes: comm_size += n.meta["tensor_meta"].numel visited_nodes.add(n) return comm_size def get_leaf(graph: Graph): """ Given a graph, return leaf nodes of this graph. Note: If we remove ``root`` nodes, ``placeholder`` nodes, and ``output`` nodes from fx graph, we will get a normal DAG. Leaf nodes in this context means leaf nodes in that DAG. """ input_nodes: Dict[Node, None] = {} for node in graph.nodes: if node.op == "output": map_arg(node.args, lambda n: input_nodes.setdefault(n)) map_arg(node.kwargs, lambda n: input_nodes.setdefault(n)) placeholder_nodes = [] for node in input_nodes.keys(): if node.op == "placeholder": placeholder_nodes.append(node) for node in placeholder_nodes: input_nodes.pop(node) return list(input_nodes.keys()) def is_leaf(graph: Graph, node: Node): return node in get_leaf(graph) def get_top(graph: Graph): """ Given a graph, return top nodes of this graph. Note: If we remove ``root`` nodes, ``placeholder`` nodes, and ``output`` nodes from fx graph, we will get a normal DAG. Top nodes in this context means nodes with BFS level 0 in that DAG. """ top_node_list = set() for node in graph.nodes: if node.op == "output": continue is_top = False def _get_top(node): nonlocal is_top if node.op == "placeholder": is_top = True map_arg(node.args, lambda n: _get_top(n)) map_arg(node.kwargs, lambda n: _get_top(n)) if is_top: top_node_list.add(node) return list(top_node_list) def is_top(graph: Graph, node: Node): return node in get_top(graph) def get_all_consumers(graph: Graph, node: Node): """ Given a graph and a node of this graph, return all consumers of the node. Returns: List of ``Nodes`` that node appear in these nodes ``args`` and ``kwargs``. """ consumer_list = [] for n in graph.nodes: if node in n.all_input_nodes: consumer_list.append(n) return consumer_list def assign_bfs_level_to_nodes(graph: Graph): """ Give a graph, assign bfs level to each node of this graph excluding ``placeholder`` and ``output`` nodes. Example: class MLP(torch.nn.Module): def __init__(self, dim: int): super().__init__() self.linear1 = torch.nn.Linear(dim, dim) self.linear2 = torch.nn.Linear(dim, dim) self.linear3 = torch.nn.Linear(dim, dim) self.linear4 = torch.nn.Linear(dim, dim) self.linear5 = torch.nn.Linear(dim, dim) def forward(self, x): l1 = self.linear1(x) l2 = self.linear2(x) l3 = self.linear3(l1) l4 = self.linear4(l2) l5 = self.linear5(l3) return l4, l5 model = MLP(4) gm = symbolic_trace(model) print(gm.graph) assign_bfs_level_to_nodes(gm.graph) for node in gm.graph.nodes: if hasattr(node, 'bfs_level'): print(node.name, node.bfs_level) Output: graph(): %x : [#users=2] = placeholder[target=x] %linear1 : [#users=1] = call_module[target=linear1](args = (%x,), kwargs = {}) %linear2 : [#users=1] = call_module[target=linear2](args = (%x,), kwargs = {}) %linear3 : [#users=1] = call_module[target=linear3](args = (%linear1,), kwargs = {}) %linear4 : [#users=1] = call_module[target=linear4](args = (%linear2,), kwargs = {}) %linear5 : [#users=1] = call_module[target=linear5](args = (%linear3,), kwargs = {}) return (linear4, linear5) linear1 0 linear2 0 linear3 1 linear4 1 linear5 2 """ current_level = 0 nodes_to_process = [] top_nodes = get_top(graph) for node in top_nodes: node.bfs_level = current_level nodes_to_process.extend(get_all_consumers(graph, node)) current_level += 1 while nodes_to_process: new_process_list = [] for node in nodes_to_process: if node.op == "output": continue node.bfs_level = current_level new_process_list.extend(get_all_consumers(graph, node)) nodes_to_process = new_process_list current_level += 1 def get_node_module(node) -> torch.nn.Module: """ Find the module associated with the given node. Args: node (torch.fx.Node): a torch.fx.Node object in the fx computation graph Returns: torch.nn.Module: the module associated with the given node """ assert ( node.graph.owning_module is not None ), "Cannot find the owning_module for node.graph, please make sure the graph is associated with a GraphModule object" assert node.op == "call_module", f"Expected node.op to be call_module, but found {node.op}" module = node.graph.owning_module.get_submodule(node.target) return module
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/meta_info_prop.py
colossalai/fx/passes/meta_info_prop.py
from dataclasses import asdict from typing import Any, Dict, List, NamedTuple, Tuple import torch import torch.fx from torch.fx.node import Argument, Node, Target from torch.utils._pytree import tree_map from colossalai.fx._compatibility import compatibility, is_compatible_with_meta from colossalai.fx.profiler import ( GraphInfo, activation_size, calculate_fwd_in, calculate_fwd_out, calculate_fwd_tmp, profile_function, profile_method, profile_module, ) @compatibility(is_backward_compatible=True) class TensorMetadata(NamedTuple): # TensorMetadata is a structure containing pertinent information # about a tensor within a PyTorch program. shape: torch.Size dtype: torch.dtype requires_grad: bool stride: Tuple[int] numel: int is_tensor: bool # TODO: we can add a list of sharding spec here, and record the sharding # behavior by appending sharding spec into list. def _extract_tensor_metadata(result: torch.Tensor) -> TensorMetadata: """ Extract a TensorMetadata NamedTuple describing `result`. """ shape = result.shape dtype = result.dtype requires_grad = result.requires_grad stride = result.stride() numel = result.numel() is_tensor = True return TensorMetadata(shape, dtype, requires_grad, stride, numel, is_tensor) @compatibility(is_backward_compatible=True) class MetaInfoProp(torch.fx.Interpreter): """ Execute an FX graph Node-by-Node with meta tensor and record the memory usage, FLOPs, and type of the result into the corresponding node. Usage: BATCH_SIZE = 2 DIM_IN = 4 DIM_HIDDEN = 16 DIM_OUT = 16 model = torch.nn.Sequential( torch.nn.Linear(DIM_IN, DIM_HIDDEN), torch.nn.Linear(DIM_HIDDEN, DIM_OUT), ) input_sample = torch.rand(BATCH_SIZE, DIM_IN) gm = symbolic_trace(model) interp = MetaInfoProp(gm) interp.run(input_sample) print(interp.summary(format='kb')) # don't panic if some statistics are 0.00 MB # output of above code is Op type Op Forward FLOPs Backward FLOPs FWD_OUT FWD_TMP BWD_OUT BWD_TMP ----------- ------- --------------- ---------------- --------- --------- --------- --------- placeholder input_1 0 FLOPs 0 FLOPs 0.00 KB 0.00 KB 0.00 KB 0.00 KB call_module _0 128 FLOPs 288 FLOPs 0.12 KB 0.00 KB 0.34 KB 0.00 KB call_module _1 512 FLOPs 1,056 FLOPs 0.12 KB 0.00 KB 1.19 KB 0.00 KB output output 0 FLOPs 0 FLOPs 0.00 KB 0.00 KB 0.00 KB 0.00 KB Args: module (GraphModule): The module to be executed """ _is_proped: bool = False @compatibility(is_backward_compatible=True) def run_node(self, n: Node) -> Any: """ Run a specific node ``n`` and return the result. Calls into placeholder, get_attr, call_function, call_method, call_module, or output depending on ``node.op`` Args: n (Node): The Node to execute Returns: Any: The result of executing ``n`` """ self._is_proped = True result, meta_info = super().run_node(n) def extract_tensor_meta(obj): if isinstance(obj, torch.Tensor): return _extract_tensor_metadata(obj) else: return TensorMetadata(None, None, False, None, 0, False) tensor_meta = tree_map(extract_tensor_meta, result) n.meta["tensor_meta"] = tensor_meta n.meta = {**n.meta, **asdict(meta_info)} # extend MetaInfo to `n.meta` # TODO: the attribute node_size should be removed in the future setattr(n, "node_size", activation_size(n.meta.get("fwd_out", 0)) + activation_size(n.meta.get("fwd_tmp", 0))) setattr(n, "fwd_flop", n.meta.get("fwd_flop", 0)) setattr(n, "bwd_flop", n.meta.get("bwd_flop", 0)) n.meta["type"] = type(result) # retain the autograd graph for param in self.module.parameters(): param.grad = None return result # Main Node running APIs @compatibility(is_backward_compatible=True) def placeholder(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``placeholder`` node. Note that this is stateful: ``Interpreter`` maintains an internal iterator over arguments passed to ``run`` and this method returns next() on that iterator. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Returns: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ return super().placeholder(target, args, kwargs), GraphInfo() @compatibility(is_backward_compatible=True) def get_attr(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``get_attr`` node. Will retrieve an attribute value from the ``Module`` hierarchy of ``self.module``. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ return super().get_attr(target, args, kwargs), GraphInfo() @compatibility(is_backward_compatible=True) def call_function(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_function`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ assert not isinstance(target, str) return profile_function(target)(*args, **kwargs) @compatibility(is_backward_compatible=True) def call_method(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_method`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ return profile_method(target)(*args, **kwargs) @compatibility(is_backward_compatible=True) def call_module(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_module`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ # Retrieve executed args and kwargs values from the environment # Execute the method and return the result assert isinstance(target, str) submod = self.fetch_attr(target) return profile_module(submod)(*args, **kwargs) @compatibility(is_backward_compatible=True) def output(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute an ``output`` node. This really just retrieves the value referenced by the ``output`` node and returns it. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ if hasattr(args[0], "_tensor"): return args[0], GraphInfo(fwd_in=[args[0]._tensor]) return args[0], GraphInfo(save_fwd_in=True) def propagate(self, *args): """ Run `module` via interpretation and return the result and record the shape and type of each node. Args: *args (Tensor): the sample input. Returns: Any: The value returned from executing the Module """ return super().run(*args) def summary(self, unit: str = "MB") -> str: """ Summarizes the memory and FLOPs statistics of the `GraphModule` in tabular format. Note that this API requires the ``tabulate`` module to be installed. """ # https://github.com/pytorch/pytorch/blob/master/torch/fx/graph.py try: from tabulate import tabulate except ImportError: print( "`summary` relies on the library `tabulate`, " "which could not be found on this machine. Run `pip " "install tabulate` to install the library." ) assert self._is_proped, "Please call `interp.run(input)` before calling `interp.summary()`." # Build up a list of summary information for each node node_summaries: List[List[Any]] = [] def mem_repr(mem: int) -> str: unit_divisor_map = { "kb": 1024, "mb": 1024**2, "gb": 1024**3, "tb": 1024**4, } return f"{mem / unit_divisor_map[unit.lower()]:.2f} {unit.upper()}" def flops_repr(flop: int) -> str: return f"{flop:,} FLOPs" accumulate_size = 0 for node in self.module.graph.nodes: node: Node accumulate_size += calculate_fwd_out(node) + calculate_fwd_tmp(node) node_summaries.append( [ node.op, str(node), flops_repr(node.meta["fwd_flop"]), flops_repr(node.meta["bwd_flop"]), mem_repr(accumulate_size), mem_repr(calculate_fwd_in(node)), mem_repr(calculate_fwd_out(node)), mem_repr(calculate_fwd_tmp(node)), mem_repr(node.meta["bwd_mem_out"]), mem_repr(node.meta["bwd_mem_tmp"]), ] ) # Use the ``tabulate`` library to create a well-formatted table # presenting our summary information headers: List[str] = [ "Op type", "Op", "Forward FLOPs", "Backward FLOPs", "Accumulated Memory", "FWD_IN", "FWD_OUT", "FWD_TMP", "BWD_OUT", "BWD_TMP", ] return tabulate(node_summaries, headers=headers, stralign="right") def metainfo_trace(gm: torch.fx.GraphModule, *args, verbose: bool = False, unit: str = "MB", **kwargs) -> None: """ MetaInfo tracing API Given a ``GraphModule`` and a sample input, this API will trace the MetaInfo of a single training cycle, and annotate them on ``gm.graph``. Uses: >>> model = ... >>> gm = symbolic_trace(model) >>> args = ... # sample input to the ``GraphModule`` >>> metainfo_trace(gm, *args) Args: gm (torch.fx.GraphModule): The ``GraphModule`` to be annotated with MetaInfo. verbose (bool, optional): Whether to show ``MetaInfoProp.summary()`. Defaults to False. unit (str, optional): The unit of memory. Defaults to "MB". Returns: torch.fx.GraphModule: The ``GraphModule`` annotated with MetaInfo. """ device = torch.device("cuda:0") if torch.cuda.is_available() else torch.device("cpu") interp = MetaInfoProp(gm.to(device)) if is_compatible_with_meta(): from colossalai.fx.profiler import MetaTensor args = tree_map(lambda x: MetaTensor(x, fake_device=device), args) kwargs = tree_map(lambda x: MetaTensor(x, fake_device=device), kwargs) interp.propagate(*args, **kwargs) if verbose: interp.summary(unit) gm.to("cpu") del interp return gm
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/concrete_info_prop.py
colossalai/fx/passes/concrete_info_prop.py
from dataclasses import asdict from typing import Any, Dict, List, Optional, Tuple import torch import torch.fx from torch.fx.node import Argument, Node, Target from torch.utils._pytree import tree_flatten from colossalai.fx._compatibility import compatibility from colossalai.fx.profiler import GraphInfo, profile_function, profile_method, profile_module @compatibility(is_backward_compatible=True) class ConcreteInfoProp(torch.fx.Interpreter): """ Execute an FX graph Node-by-Node with concrete tensor and record the memory usage, execution time of forward and backward, and type of the result into the corresponding node. Usage: BATCH_SIZE = 2 DIM_IN = 4 DIM_HIDDEN = 16 DIM_OUT = 16 model = torch.nn.Sequential( torch.nn.Linear(DIM_IN, DIM_HIDDEN), torch.nn.Linear(DIM_HIDDEN, DIM_OUT), ).cuda() input_sample = torch.rand(BATCH_SIZE, DIM_IN, device="cuda") gm = symbolic_trace(model) interp = ConcreteInfoProp(gm) interp.run(input_sample) print(interp.summary(unit='kb')) output of above code is Op type Op Forward time Backward time SAVE_FWD_IN FWD_OUT FWD_TMP BWD_OUT BWD_TMP ----------- ------- ----------------------- ------------------------ ------------- --------- --------- --------- --------- placeholder input_1 0.0 s 0.0 s False 0.00 KB 0.00 KB 0.00 KB 0.00 KB call_module _0 0.0003993511199951172 s 0.00706791877746582 s False 0.50 KB 0.00 KB 0.03 KB 0.66 KB call_module _1 6.29425048828125e-05 s 0.00018286705017089844 s False 0.50 KB 0.00 KB 0.12 KB 0.81 KB output output 0.0 s 0.0 s True 0.00 KB 0.00 KB 0.00 KB 0.00 KB Args: module (GraphModule): The module to be executed """ _is_proped: bool = False def run(self, *args, initial_env: Optional[Dict[Node, Any]] = None, enable_io_processing: bool = True) -> Any: """Customized run for ConcreteInfoProp We need to store the device in self.device Args: *args: The arguments to the Module to run, in positional order initial_env (Optional[Dict[Node, Any]]): An optional starting environment for execution. This is a dict mapping `Node` to any value. This can be used, for example, to pre-populate results for certain `Nodes` so as to do only partial evaluation within the interpreter. enable_io_processing (bool): If true, we process the inputs and outputs with graph's process_inputs and process_outputs function first before using them. Returns: Any: The value returned from executing the Module """ flatten_args, _ = tree_flatten(args) self.device = next(item for item in flatten_args if hasattr(item, "device")).device return super().run(*args, initial_env, enable_io_processing) @compatibility(is_backward_compatible=True) def run_node(self, n: Node) -> Any: """ Run a specific node ``n`` and return the result. Calls into placeholder, get_attr, call_function, call_method, call_module, or output depending on ``node.op`` Args: n (Node): The Node to execute Returns: Any: The result of executing ``n`` """ self._is_proped = True result, meta_info = super().run_node(n) n.meta = {**n.meta, **asdict(meta_info)} # extend MetaInfo to `n.meta` # TODO: the attribute node_size should be removed in the future setattr(n, "node_size", n.meta.get("fwd_mem_tmp", 0) + n.meta.get("fwd_mem_out", 0)) n.meta["type"] = type(result) # retain the autograd graph for param in self.module.parameters(): param.grad = None return result # Main Node running APIs @compatibility(is_backward_compatible=True) def placeholder(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``placeholder`` node. Note that this is stateful: ``Interpreter`` maintains an internal iterator over arguments passed to ``run`` and this method returns next() on that iterator. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Returns: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and forward & backward time. """ return super().placeholder(target, args, kwargs), GraphInfo() @compatibility(is_backward_compatible=True) def get_attr(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``get_attr`` node. Will retrieve an attribute value from the ``Module`` hierarchy of ``self.module``. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and FLOPs estimated with `MetaTensor`. """ return super().get_attr(target, args, kwargs), GraphInfo() @compatibility(is_backward_compatible=True) def call_function(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_function`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and forward & backward time. """ assert not isinstance(target, str) return profile_function(target, self.device)(*args, **kwargs) @compatibility(is_backward_compatible=True) def call_method(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_method`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and forward & backward time. """ return profile_method(target, self.device)(*args, **kwargs) @compatibility(is_backward_compatible=True) def call_module(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute a ``call_module`` node with meta tensor and return the result and its meta profile. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and forward & backward time. """ # Retrieve executed args and kwargs values from the environment # Execute the method and return the result assert isinstance(target, str) submod = self.fetch_attr(target) return profile_module(submod, self.device)(*args, **kwargs) @compatibility(is_backward_compatible=True) def output(self, target: "Target", args: Tuple[Argument, ...], kwargs: Dict[str, Any]) -> Any: """ Execute an ``output`` node. This really just retrieves the value referenced by the ``output`` node and returns it. Args: target (Target): The call target for this node. See `Node <https://pytorch.org/docs/master/fx.html#torch.fx.Node>`__ for details on semantics args (Tuple): Tuple of positional args for this invocation kwargs (Dict): Dict of keyword arguments for this invocation Return: result (Any): The argument value that was retrieved meta_info (MetaInfo): The memory cost and forward & backward time. """ return args[0], GraphInfo(save_fwd_in=True) def propagate(self, *args): """ Run `module` via interpretation and return the result and record the shape and type of each node. Args: *args (Tensor): the sample input. Returns: Any: The value returned from executing the Module """ return self.run(*args) def summary(self, unit: str = "MB") -> str: """ Summarizes the memory and FLOPs statistics of the `GraphModule` in tabular format. Note that this API requires the ``tabulate`` module to be installed. """ # https://github.com/pytorch/pytorch/blob/master/torch/fx/graph.py try: from tabulate import tabulate except ImportError: print( "`summary` relies on the library `tabulate`, " "which could not be found on this machine. Run `pip " "install tabulate` to install the library." ) assert self._is_proped, "Please call `interp.run(input)` before calling `interp.summary()`." # Build up a list of summary information for each node node_summaries: List[List[Any]] = [] def mem_repr(mem: int) -> str: unit_divisor_map = { "kb": 1024, "mb": 1024**2, "gb": 1024**3, "tb": 1024**4, } return f"{mem / unit_divisor_map[unit.lower()]:.2f} {unit.upper()}" def time_repr(time: float): return f"{time:,} s" for node in self.module.graph.nodes: node: Node node_summaries.append( [ node.op, str(node), time_repr(node.meta["fwd_time"]), time_repr(node.meta["bwd_time"]), node.meta["save_fwd_in"], mem_repr(node.meta["fwd_mem_out"]), mem_repr(node.meta["fwd_mem_tmp"]), mem_repr(node.meta["bwd_mem_out"]), mem_repr(node.meta["bwd_mem_tmp"]), ] ) # Use the ``tabulate`` library to create a well-formatted table # presenting our summary information headers: List[str] = [ "Op type", "Op", "Forward time", "Backward time", "SAVE_FWD_IN", "FWD_OUT", "FWD_TMP", "BWD_OUT", "BWD_TMP", ] return tabulate(node_summaries, headers=headers, stralign="right")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/__init__.py
colossalai/fx/passes/__init__.py
from .adding_split_node_pass import balanced_split_pass, split_with_split_nodes_pass from .concrete_info_prop import ConcreteInfoProp from .meta_info_prop import MetaInfoProp, metainfo_trace from .shard_1d_pass import column_shard_linear_pass, row_shard_linear_pass
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/split_module.py
colossalai/fx/passes/split_module.py
import inspect from typing import Any, Callable, Dict, List, Optional import torch from packaging import version from torch.fx._compatibility import compatibility from torch.fx.graph_module import GraphModule @compatibility(is_backward_compatible=True) class Partition: """ Adapted from https://github.com/pytorch/pytorch/blob/master/torch/fx/passes/split_module.py """ def __init__(self, name: str): self.name: str = name self.node_names: List[str] = [] self.inputs: Dict[str, None] = {} self.outputs: Dict[str, None] = {} self.partitions_dependent_on: Dict[str, None] = {} self.partition_dependents: Dict[str, None] = {} self.graph: torch.fx.graph.Graph = torch.fx.graph.Graph() self.environment: Dict[torch.fx.node.Node, torch.fx.node.Node] = {} self.targets: Dict[str, Any] = {} def __repr__(self) -> str: return ( f"name: {self.name},\n" f" nodes: {self.node_names},\n" f" inputs: {self.inputs},\n" f" outputs: {self.outputs},\n" f" partitions dependent on: {self.partitions_dependent_on},\n" f" partition dependents: {self.partition_dependents}" ) # Creates subgraphs out of main graph @compatibility(is_backward_compatible=True) def split_module( m: GraphModule, root_m: torch.nn.Module, split_callback: Callable[[torch.fx.node.Node], int], merge_output=False, ): """ Adapted from https://github.com/pytorch/pytorch/blob/master/torch/fx/passes/split_module.py Creates subgraphs out of main graph Args: m (GraphModule): Graph module to split root_m (torch.nn.Module): root nn module. Not currently used. Included because the root nn module is usually transformed via torch.fx._symbolic_trace.symbolic_trace (see example below) split_callback (Callable[[torch.fx.node.Node], int]): Callable function that maps a given Node instance to a numeric partition identifier. split_module will use this function as the policy for which operations appear in which partitions in the output Module. Returns: GraphModule: the module after split. Example: This is a sample setup: import torch from torch.fx.symbolic_trace import symbolic_trace from torch.fx.graph_module import GraphModule from torch.fx.node import Node from colossalai.fx.passes.split_module import split_module class MyModule(torch.nn.Module): def __init__(self): super().__init__() self.param = torch.nn.Parameter(torch.rand(3, 4)) self.linear = torch.nn.Linear(4, 5) def forward(self, x, y): z = self.linear(x + self.param).clamp(min=0.0, max=1.0) w = self.linear(y).clamp(min=0.0, max=1.0) return z + w # symbolically trace model my_module = MyModule() my_module_traced = symbolic_trace(my_module) # random mod partitioning partition_counter = 0 NPARTITIONS = 3 def mod_partition(node: Node): global partition_counter partition = partition_counter % NPARTITIONS partition_counter = (partition_counter + 1) % NPARTITIONS return partition # split module in module with submodules module_with_submodules = split_module( my_module_traced, my_module, mod_partition ) Output looks like this. Original graph is broken into partitions > print(module_with_submodules) GraphModule( (submod_0): GraphModule( (linear): Linear(in_features=4, out_features=5, bias=True) ) (submod_1): GraphModule( (linear): Linear(in_features=4, out_features=5, bias=True) ) (submod_2): GraphModule() ) def forward(self, x, y): param = self.param submod_0 = self.submod_0(x, param, y); x = param = y = None getitem = submod_0[0] getitem_1 = submod_0[1]; submod_0 = None submod_1 = self.submod_1(getitem, getitem_1); getitem = getitem_1 = None getitem_2 = submod_1[0] getitem_3 = submod_1[1]; submod_1 = None submod_2 = self.submod_2(getitem_2, getitem_3); getitem_2 = getitem_3 = None return submod_2 Output of split module is the same as output of input traced module. This is an example within a test setting: > orig_out = my_module_traced(x, y) > submodules_out = module_with_submodules(x, y) > self.assertEqual(orig_out, submodules_out) True """ partitions: Dict[str, Partition] = {} orig_nodes: Dict[str, torch.fx.node.Node] = {} def record_cross_partition_use(def_node: torch.fx.node.Node, use_node: Optional[torch.fx.node.Node]): # noqa: B950 def_partition_name = getattr(def_node, "_fx_partition", None) use_partition_name = getattr(use_node, "_fx_partition", None) if def_partition_name != use_partition_name: if def_partition_name is not None: def_partition = partitions[def_partition_name] def_partition.outputs.setdefault(def_node.name) if use_partition_name is not None: def_partition.partition_dependents.setdefault(use_partition_name) if use_partition_name is not None: use_partition = partitions[use_partition_name] use_partition.inputs.setdefault(def_node.name) if def_partition_name is not None: use_partition.partitions_dependent_on.setdefault(def_partition_name) def record_output(def_node: torch.fx.node.Node, use_node: Optional[torch.fx.node.Node]): # noqa: B950 def_partition_name = getattr(def_node, "_fx_partition", None) use_partition_name = getattr(use_node, "_fx_partition", None) if def_partition_name != use_partition_name: if def_partition_name is not None: def_partition = partitions[def_partition_name] def_partition.outputs.setdefault(def_node.name) if use_partition_name is not None: def_partition.partition_dependents.setdefault(use_partition_name) if use_partition_name is not None: use_partition = partitions[use_partition_name] use_partition.inputs.setdefault(def_node.name) if def_partition_name is not None: use_partition.partitions_dependent_on.setdefault(def_partition_name) use_partition.outputs.setdefault(def_node.name) else: if use_partition_name is not None: use_partition = partitions[use_partition_name] use_partition.outputs.setdefault(def_node.name) # split nodes into partitions for node in m.graph.nodes: orig_nodes[node.name] = node if node.op in ["placeholder"]: continue if node.op == "output": if merge_output: torch.fx.graph.map_arg(node.args[0], lambda n: record_output(n, node.prev)) else: torch.fx.graph.map_arg(node.args[0], lambda n: record_cross_partition_use(n, None)) continue partition_name = str(split_callback(node)) # add node to partitions partition = partitions.get(partition_name) if partition is None: partitions[partition_name] = partition = Partition(partition_name) partition.node_names.append(node.name) node._fx_partition = partition_name torch.fx.graph.map_arg(node.args, lambda def_node: record_cross_partition_use(def_node, node)) torch.fx.graph.map_arg(node.kwargs, lambda def_node: record_cross_partition_use(def_node, node)) # noqa: B950 # find partitions with no dependencies root_partitions: List[str] = [] for partition_name, partition in partitions.items(): if not len(partition.partitions_dependent_on): root_partitions.append(partition_name) # check partitions for circular dependencies and create topological partition ordering sorted_partitions: List[str] = [] while root_partitions: root_partition = root_partitions.pop() sorted_partitions.append(root_partition) for dependent in partitions[root_partition].partition_dependents: partitions[dependent].partitions_dependent_on.pop(root_partition) if not partitions[dependent].partitions_dependent_on: root_partitions.append(dependent) if len(sorted_partitions) != len(partitions): raise RuntimeError("cycle exists between partitions!") # add placeholders to partitions for partition_name in sorted_partitions: partition = partitions[partition_name] for input in partition.inputs: placeholder = partition.graph.placeholder(input) placeholder.meta = orig_nodes[input].meta.copy() partition.environment[orig_nodes[input]] = placeholder # Transform nodes and collect targets for partition's submodule for node in m.graph.nodes: if hasattr(node, "_fx_partition"): partition = partitions[node._fx_partition] # swap out old graph nodes in kw/args with references to new nodes in this submodule environment = partition.environment gathered_args = torch.fx.graph.map_arg(node.args, lambda n: environment[n]) gathered_kwargs = torch.fx.graph.map_arg(node.kwargs, lambda n: environment[n]) if node.op not in ["call_module", "get_attr"]: target = node.target else: target_atoms = node.target.split(".") target_attr = m for atom in target_atoms: if not hasattr(target_attr, atom): raise RuntimeError(f"Operator target {node.target} not found!") target_attr = getattr(target_attr, atom) # target = target_atoms[-1] target = "_".join(target_atoms) partition.targets[target] = target_attr assert isinstance(gathered_args, tuple) assert isinstance(gathered_kwargs, dict) new_node = partition.graph.create_node( op=node.op, target=target, args=gathered_args, kwargs=gathered_kwargs ) new_node.meta = node.meta.copy() partition.environment[node] = new_node # Set up values to construct base module base_mod_env: Dict[str, torch.fx.node.Node] = {} base_mod_graph: torch.fx.graph.Graph = torch.fx.graph.Graph() base_mod_attrs: Dict[str, torch.fx.graph_module.GraphModule] = {} for node in m.graph.nodes: if node.op == "placeholder": if version.parse(torch.__version__) < version.parse("1.11.0"): base_mod_env[node.name] = base_mod_graph.placeholder(node.target, type_expr=node.type) else: default_value = node.args[0] if len(node.args) > 0 else inspect.Signature.empty base_mod_env[node.name] = base_mod_graph.placeholder( node.target, type_expr=node.type, default_value=default_value ) base_mod_env[node.name].meta = node.meta.copy() # Do some things iterating over the partitions in topological order again: # 1) Finish off submodule Graphs by setting corresponding outputs # 2) Construct GraphModules for each submodule # 3) Construct the base graph by emitting calls to those submodules in # topological order for partition_name in sorted_partitions: partition = partitions[partition_name] # Set correct output values output_vals = tuple(partition.environment[orig_nodes[name]] for name in partition.outputs) output_vals = output_vals[0] if len(output_vals) == 1 else output_vals # type: ignore[assignment] partition.graph.output(output_vals) # Construct GraphModule for this partition submod_name = f"submod_{partition_name}" base_mod_attrs[submod_name] = torch.fx.graph_module.GraphModule( partition.targets, partition.graph ) # noqa: B950 # Emit call in base graph to this submodule output_val = base_mod_graph.call_module(submod_name, tuple(base_mod_env[name] for name in partition.inputs)) if len(partition.outputs) > 1: # Unpack multiple return values from submodule output_val_proxy = torch.fx.proxy.Proxy(output_val) for i, output_name in enumerate(partition.outputs): base_mod_env[output_name] = output_val_proxy[i].node # type: ignore[index] else: if not partition.outputs: continue base_mod_env[list(partition.outputs)[0]] = output_val for node in m.graph.nodes: if node.op == "output": base_mod_graph.output(torch.fx.graph.map_arg(node.args[0], lambda n: base_mod_env[n.name])) # noqa: B950 for partition_name in sorted_partitions: partition = partitions[partition_name] new_gm = torch.fx.graph_module.GraphModule(base_mod_attrs, base_mod_graph) return new_gm
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false