repo
stringlengths
7
90
file_url
stringlengths
81
315
file_path
stringlengths
4
228
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 14:38:15
2026-01-05 02:33:18
truncated
bool
2 classes
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/passes_for_gpt2_test.py
colossalai/fx/passes/passes_for_gpt2_test.py
import inspect from typing import Any, Callable, Dict, List, Optional import torch from packaging import version from torch.fx._compatibility import compatibility from torch.fx.graph_module import GraphModule from colossalai.fx.passes.adding_split_node_pass import balanced_split_pass, pipe_split from colossalai.fx.passes.meta_info_prop import TensorMetadata from colossalai.fx.passes.split_module import Partition def customized_split_pass_for_gpt2(gm: torch.fx.GraphModule, pp_size: int, partition_list: List[int]): """ This pass is only used to do the gpt2 performance test, it may move into adding_split_node_pass.py, and will be deprecated in future. """ mod_graph = gm.graph valid_children_size = 0 valid_children = [] for node in mod_graph.nodes: if node.op == "call_module": valid_children_size += 1 valid_children.append(node.target) if valid_children_size < pp_size: # If valid children is not enough to shard, we will use balanced policy instead of uniform policy. return balanced_split_pass(gm, pp_size) accumulate_layer_amount = 0 list_of_part = partition_list part_index = 0 for node in mod_graph.nodes: if pp_size <= 1: break if node.op == "call_module": if node.target in valid_children: accumulate_layer_amount += 1 if accumulate_layer_amount == list_of_part[part_index]: part_index += 1 pp_size -= 1 with mod_graph.inserting_after(node): split_node = mod_graph.create_node("call_function", pipe_split) gm.recompile() return gm def split_with_split_nodes_pass_for_gp2_test(annotated_gm: torch.fx.GraphModule): """ This pass will be used in gpt2 test, only a part of changes may be added into split_with_split_nodes_pass, and it will be deprecated in future. """ part_idx = 0 def eliminate_unused_placeholders(gm): for node in gm.graph.nodes: if node.op == "placeholder": if not len(node.users): gm.graph.erase_node(node) gm.recompile() return gm def refill_outputs_and_placeholders(gm, next_partition_placeholders): """ This method is used to eliminate the outputs in previous partition which is unused in next partition. In split module pass, it treats partitions as a DAG, but we need treat them as a single direction linked list in pipeline parallel. The difference is if a output from partition 0 is an input argument of partition 3, the DAG will not transfer it to partition 1 and partition 2. However, in single direction linked list, we need to do so. """ output_type = None output_args = [] non_output_list = [] new_placeholder_list = [] for node in gm.graph.nodes: if node.op == "output": if isinstance(node.args[0], (tuple, list)): output_type = node.args[0].__class__ output_args.extend([n.name for n in node.args[0]]) else: output_args.append(node.args[0].name) rm_list = [] for name in output_args: if next_partition_placeholders and name not in next_partition_placeholders: rm_list.append(name) for name in rm_list: output_args.remove(name) gm.graph.erase_node(node) else: non_output_list.append(node.name) for name in next_partition_placeholders: if name not in output_args: output_args.append(name) for name in output_args: if name not in non_output_list: gm.graph.placeholder(name) # convert name to node for output_args for index, name in enumerate(output_args): for n in gm.graph.nodes: if n.name == name: output_args[index] = n continue # reorder the output args to make sure # output args has same order as next partition placeholder reorder_output_args = [] if next_partition_placeholders: for name in next_partition_placeholders: for node in output_args: if node.name == name: reorder_output_args.append(node) continue for node in gm.graph.nodes: if node.op == "placeholder": new_placeholder_list.append(node.name) if output_type is not None: gm.graph.output(output_type(output_args)) else: gm.graph.output(output_args) gm.recompile() return gm, new_placeholder_list def split_callback(n: torch.fx.Node): nonlocal part_idx if (n.op, n.target) == ("call_function", pipe_split): part_idx += 1 return part_idx split_mod = split_module_for_gpt2_test(annotated_gm, None, split_callback) split_submodules = [] for name, submodule in split_mod.named_modules(): if isinstance(submodule, torch.fx.GraphModule): for node in submodule.graph.nodes: if (node.op, node.target) == ("call_function", pipe_split): submodule.graph.erase_node(node) submodule.recompile() split_submodules.append(submodule) submodules = list(split_mod.children()) placeholder_dict = {} for submodule in submodules: submodule = eliminate_unused_placeholders(submodule) placeholder_dict[submodule] = [] submodules.reverse() for index, submodule in enumerate(submodules): if index == 0: placeholder_list = [] else: placeholder_list = placeholder_dict[submodules[index - 1]] submodule, placeholder_dict[submodule] = refill_outputs_and_placeholders(submodule, placeholder_list) submodule.recompile() split_mod.recompile() return split_mod, split_submodules @compatibility(is_backward_compatible=True) def split_module_for_gpt2_test( m: GraphModule, root_m: torch.nn.Module, split_callback: Callable[[torch.fx.node.Node], int], ): """ This pass will be used in gpt2 pp performance test, only a part of changes may be added into split_module, and it will be deprecated in future. """ partitions: Dict[str, Partition] = {} orig_nodes: Dict[str, torch.fx.node.Node] = {} def _node_with_all_tensor_element(node_metadata: Any) -> int: """ return whether node contains non-tensor element. """ all_tensor_node = True if isinstance(node_metadata, TensorMetadata): all_tensor_node = node_metadata.is_tensor and all_tensor_node elif isinstance(node_metadata, dict): value_list = [v for _, v in node_metadata.items()] all_tensor_node += _node_with_all_tensor_element(value_list) else: for element in node_metadata: all_tensor_node += _node_with_all_tensor_element(element) return all_tensor_node def _move_all_ancestors_into_partition(node, partition_name): all_ancestors = set() def _gen_all_ancestors_set(node): all_ancestors.add(node) for n in node.all_input_nodes: if n in all_ancestors: continue _gen_all_ancestors_set(n) _gen_all_ancestors_set(node) for n in list(all_ancestors): if n.op != "placeholder" and n._fx_partition > partition_name: n._fx_partition = partition_name def record_cross_partition_use(def_node: torch.fx.node.Node, use_node: Optional[torch.fx.node.Node]): # noqa: B950 def_partition_name = getattr(def_node, "_fx_partition", None) use_partition_name = getattr(use_node, "_fx_partition", None) if def_partition_name != use_partition_name: # if 'tensor_meta' in def_node.meta: # if not _node_with_all_tensor_element(def_node.meta['tensor_meta']): # _move_all_ancestors_into_partition(use_node, def_partition_name) # node_process_list.extend(use_node.all_input_nodes) # node_process_list.extend(list(use_node.users)) # node_process_list.append(use_node) # return if def_partition_name is not None: def_partition = partitions[def_partition_name] def_partition.outputs.setdefault(def_node.name) if use_partition_name is not None: def_partition.partition_dependents.setdefault(use_partition_name) if use_partition_name is not None: use_partition = partitions[use_partition_name] use_partition.inputs.setdefault(def_node.name) if def_partition_name is not None: use_partition.partitions_dependent_on.setdefault(def_partition_name) node_process_list = list(m.graph.nodes) # split nodes into partitions while node_process_list: node = node_process_list.pop(0) orig_nodes[node.name] = node if node.op in ["placeholder"]: continue if node.op == "output": # partition_name = str(split_callback(node)) # def _set_output_args_partition(n, partition_name): # n._fx_partition = partition_name # torch.fx.graph.map_arg(node.args[0], lambda n: _set_output_args_partition(n, partition_name)) torch.fx.graph.map_arg(node.args[0], lambda n: record_cross_partition_use(n, None)) continue partition_name = str(split_callback(node)) # add node to partitions partition = partitions.get(partition_name) if partition is None: partitions[partition_name] = partition = Partition(partition_name) partition.node_names.append(node.name) origin_partition_name = getattr(node, "_fx_partition", None) if origin_partition_name is None: node._fx_partition = partition_name torch.fx.graph.map_arg(node.args, lambda def_node: record_cross_partition_use(def_node, node)) torch.fx.graph.map_arg(node.kwargs, lambda def_node: record_cross_partition_use(def_node, node)) # noqa: B950 # find partitions with no dependencies root_partitions: List[str] = [] for partition_name, partition in partitions.items(): if not len(partition.partitions_dependent_on): root_partitions.append(partition_name) # check partitions for circular dependencies and create topological partition ordering sorted_partitions: List[str] = [] while root_partitions: root_partition = root_partitions.pop() sorted_partitions.append(root_partition) for dependent in partitions[root_partition].partition_dependents: partitions[dependent].partitions_dependent_on.pop(root_partition) if not partitions[dependent].partitions_dependent_on: root_partitions.append(dependent) if len(sorted_partitions) != len(partitions): raise RuntimeError("cycle exists between partitions!") # add placeholders to partitions for partition_name in sorted_partitions: partition = partitions[partition_name] for input in partition.inputs: placeholder = partition.graph.placeholder(input) placeholder.meta = orig_nodes[input].meta.copy() partition.environment[orig_nodes[input]] = placeholder # Transform nodes and collect targets for partition's submodule for node in m.graph.nodes: if hasattr(node, "_fx_partition"): partition = partitions[node._fx_partition] # swap out old graph nodes in kw/args with references to new nodes in this submodule environment = partition.environment gathered_args = torch.fx.graph.map_arg(node.args, lambda n: environment[n]) gathered_kwargs = torch.fx.graph.map_arg(node.kwargs, lambda n: environment[n]) if node.op not in ["call_module", "get_attr"]: target = node.target else: target_atoms = node.target.split(".") target_attr = m for atom in target_atoms: if not hasattr(target_attr, atom): raise RuntimeError(f"Operator target {node.target} not found!") target_attr = getattr(target_attr, atom) # target = target_atoms[-1] target = "_".join(target_atoms) partition.targets[target] = target_attr assert isinstance(gathered_args, tuple) assert isinstance(gathered_kwargs, dict) new_node = partition.graph.create_node( op=node.op, target=target, args=gathered_args, kwargs=gathered_kwargs, name=node.name ) new_node.meta = node.meta.copy() partition.environment[node] = new_node # Set up values to construct base module base_mod_env: Dict[str, torch.fx.node.Node] = {} base_mod_graph: torch.fx.graph.Graph = torch.fx.graph.Graph() base_mod_attrs: Dict[str, torch.fx.graph_module.GraphModule] = {} for node in m.graph.nodes: if node.op == "placeholder": if version.parse(torch.__version__) < version.parse("1.11.0"): base_mod_env[node.name] = base_mod_graph.placeholder(node.name, type_expr=node.type) else: default_value = node.args[0] if len(node.args) > 0 else inspect.Signature.empty base_mod_env[node.name] = base_mod_graph.placeholder( node.name, type_expr=node.type, default_value=default_value ) base_mod_env[node.name].meta = node.meta.copy() # Do some things iterating over the partitions in topological order again: # 1) Finish off submodule Graphs by setting corresponding outputs # 2) Construct GraphModules for each submodule # 3) Construct the base graph by emitting calls to those submodules in # topological order for partition_name in sorted_partitions: partition = partitions[partition_name] # Set correct output values output_vals = tuple(partition.environment[orig_nodes[name]] for name in partition.outputs) output_vals = output_vals[0] if len(output_vals) == 1 else output_vals # type: ignore[assignment] partition.graph.output(output_vals) # Construct GraphModule for this partition submod_name = f"submod_{partition_name}" base_mod_attrs[submod_name] = torch.fx.graph_module.GraphModule( partition.targets, partition.graph ) # noqa: B950 # Emit call in base graph to this submodule output_val = base_mod_graph.call_module(submod_name, tuple(base_mod_env[name] for name in partition.inputs)) if len(partition.outputs) > 1: # Unpack multiple return values from submodule output_val_proxy = torch.fx.proxy.Proxy(output_val) for i, output_name in enumerate(partition.outputs): base_mod_env[output_name] = output_val_proxy[i].node # type: ignore[index] else: if not partition.outputs: continue base_mod_env[list(partition.outputs)[0]] = output_val for node in m.graph.nodes: if node.op == "output": base_mod_graph.output(torch.fx.graph.map_arg(node.args[0], lambda n: base_mod_env[n.name])) # noqa: B950 return torch.fx.graph_module.GraphModule(base_mod_attrs, base_mod_graph)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/fx/passes/experimental/adding_shape_consistency_pass.py
colossalai/fx/passes/experimental/adding_shape_consistency_pass.py
import builtins import operator from typing import List import torch from colossalai.tensor.shape_consistency import ShapeConsistencyManager from colossalai.tensor.sharding_spec import ShardingSpec def apply(*args, **kwargs): shape_consistency_manager = ShapeConsistencyManager() return shape_consistency_manager.apply(*args, **kwargs) def solution_annotation_pass(gm: torch.fx.GraphModule, solution: List[int], device_mesh): mod_graph = gm.graph nodes = tuple(mod_graph.nodes) # the dict to get origin sharding spec of node origin_node_sharding_spec_dict = {} for node_index, (node, strategy_index) in enumerate(zip(nodes, solution)): strategies_vector = node.strategies_vector setattr(node, "best_strategy", strategies_vector[strategy_index]) setattr(node, "sharding_spec", strategies_vector[strategy_index].output_sharding_spec) origin_node_sharding_spec_dict[node_index] = strategies_vector[strategy_index].output_sharding_spec # apply the sharding spec of parameters for node in nodes: if node.op == "call_module": target_module = node.graph.owning_module.get_submodule(node.target) origin_sharding_spec = ShardingSpec(device_mesh, target_module.weight.shape, {}) setattr(target_module.weight, "sharding_spec", origin_sharding_spec) target_weight_sharding_spec = node.best_strategy.input_shardings[1] target_module.weight.data = target_module.weight.data.permute((1, 0, 2, 3)) apply(target_module.weight, target_weight_sharding_spec) target_module.weight.data = target_module.weight.data.permute((1, 0, 2, 3)) # the dict to get input sharding specs of user node sharding_spec_convert_dict = {} for index, node in enumerate(nodes): target_sharding_specs = [] for user_node in node.strategies_vector.successor_nodes: node_index = user_node.strategies_vector.predecessor_nodes.index(node) target_sharding_spec = user_node.best_strategy.input_shardings[node_index] target_sharding_specs.append(target_sharding_spec) sharding_spec_convert_dict[index] = target_sharding_specs # add above dicts into graph for node in nodes: if node.op != "placeholder": with mod_graph.inserting_before(node): input_specs_node = mod_graph.create_node("placeholder", target="sharding_spec_convert_dict") origin_specs_node = mod_graph.create_node("placeholder", target="origin_node_sharding_spec_dict") break return sharding_spec_convert_dict, origin_node_sharding_spec_dict def shape_consistency_pass(gm: torch.fx.GraphModule): mod_graph = gm.graph nodes = tuple(mod_graph.nodes) input_dict_node = None origin_dict_node = None # mapping the node into the origin graph index node_to_index_dict = {} index = 0 for node in nodes: if node.target == "sharding_spec_convert_dict": input_dict_node = node continue if node.target == "origin_node_sharding_spec_dict": origin_dict_node = node continue if not hasattr(node, "best_strategy"): continue node_to_index_dict[node] = index index += 1 assert input_dict_node is not None # add shape consistency apply function into graph for node in nodes: if not hasattr(node, "best_strategy"): continue with mod_graph.inserting_after(node): origin_spec_node = mod_graph.create_node( "call_function", operator.getitem, args=(origin_dict_node, node_to_index_dict[node]) ) with mod_graph.inserting_after(origin_spec_node): set_sharding_spec_node = mod_graph.create_node( "call_function", builtins.setattr, args=(node, "sharding_spec", origin_spec_node) ) for user_node in node.strategies_vector.successor_nodes: node_index = user_node.strategies_vector.predecessor_nodes.index(node) with mod_graph.inserting_before(user_node): input_specs_node = mod_graph.create_node( "call_function", operator.getitem, args=(input_dict_node, node_to_index_dict[node]) ) with mod_graph.inserting_before(user_node): sharding_spec_node = mod_graph.create_node( "call_function", operator.getitem, args=(input_specs_node, node_index) ) with mod_graph.inserting_before(user_node): shape_consistency_node = mod_graph.create_node("call_function", apply, args=(node, sharding_spec_node)) return gm
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/moe/_operation.py
colossalai/moe/_operation.py
from typing import Any, List, Optional, Tuple import torch import torch.distributed as dist from torch import Tensor from torch.cuda.amp import custom_bwd, custom_fwd from torch.distributed import ProcessGroup from colossalai.quantization.fp8 import all_to_all_single_fp8 MOE_KERNEL = None def load_moe(): global MOE_KERNEL from colossalai.kernel.kernel_loader import MoeLoader MOE_KERNEL = MoeLoader().load() class AllGather(torch.autograd.Function): @staticmethod def forward( ctx: Any, inputs: Tensor, group: Optional[ProcessGroup] = None, overlap: bool = False, ) -> Tuple[Tensor, Any]: """ Returns: outputs: Tensor handle: Optional[Work], if overlap is True """ assert ctx is not None or not overlap if ctx is not None: ctx.comm_grp = group comm_size = dist.get_world_size(group) if comm_size == 1: return inputs.unsqueeze(0), None buffer_shape = (comm_size,) + inputs.shape outputs = torch.empty(buffer_shape, dtype=inputs.dtype, device=inputs.device) buffer_list = list(torch.chunk(outputs, comm_size, dim=0)) if not overlap: dist.all_gather(buffer_list, inputs, group=group) return outputs, None else: handle = dist.all_gather(buffer_list, inputs, group=group, async_op=True) return outputs, handle @staticmethod def backward(ctx: Any, *grad_outputs) -> Tuple[Tensor, None, None]: return ( ReduceScatter.forward(None, grad_outputs[0], ctx.comm_grp, False)[0], None, None, ) class ReduceScatter(torch.autograd.Function): @staticmethod def forward( ctx: Any, inputs: Tensor, group: ProcessGroup, overlap: bool = False, ) -> Tuple[Tensor, Any]: """ Returns: outputs: Tensor handle: Optional[Work], if overlap is True """ assert ctx is not None or not overlap if ctx is not None: ctx.comm_grp = group comm_size = dist.get_world_size(group) if comm_size == 1: return inputs.squeeze(0), None if not inputs.is_contiguous(): inputs = inputs.contiguous() output_shape = inputs.shape[1:] outputs = torch.empty(output_shape, dtype=inputs.dtype, device=inputs.device) buffer_list = list(torch.chunk(inputs, comm_size, dim=0)) if not overlap: dist.reduce_scatter(outputs, buffer_list, group=group) return outputs, None else: handle = dist.reduce_scatter(outputs, buffer_list, group=group, async_op=True) return outputs, handle @staticmethod def backward(ctx: Any, *grad_outputs) -> Tuple[Tensor, None, None]: # TODO: support async backward return ( AllGather.forward(None, grad_outputs[0], ctx.comm_grp, False)[0], None, None, ) class AllToAll(torch.autograd.Function): """Dispatches input tensor [e, c, h] to all experts by all_to_all_single operation in torch.distributed. """ @staticmethod def forward( ctx: Any, inputs: Tensor, group: ProcessGroup, overlap: bool = False, ) -> Tuple[Tensor, Any]: """ Returns: outputs: Tensor handle: Optional[Work], if overlap is True """ assert ctx is not None or not overlap if ctx is not None: ctx.comm_grp = group if not inputs.is_contiguous(): inputs = inputs.contiguous() if dist.get_world_size(group) == 1: return inputs, None output = torch.empty_like(inputs) if not overlap: dist.all_to_all_single(output, inputs, group=group) return output, None else: handle = dist.all_to_all_single(output, inputs, group=group, async_op=True) return output, handle @staticmethod def backward(ctx: Any, *grad_outputs) -> Tuple[Tensor, None, None]: return ( AllToAll.forward(None, grad_outputs[0], ctx.comm_grp, False)[0], None, None, ) class HierarchicalAllToAll(torch.autograd.Function): @staticmethod def forward(ctx: Any, inputs: Tensor, groups: Tuple[ProcessGroup, ProcessGroup], src_rank: int) -> Tensor: """ Returns: outputs: Tensor """ # TODO: we can reduce comm volume by removing empty capacity if ctx is not None: ctx.comm_grps = groups ctx.src_rank = src_rank intra_node_group, inter_node_group = groups local_world_size = dist.get_world_size(intra_node_group) num_group = dist.get_world_size(inter_node_group) if inter_node_group is not None else 1 world_size = local_world_size * num_group outputs = torch.empty_like(inputs) if dist.get_rank() == src_rank: # intra-node gather intra_output = [torch.empty_like(inputs) for _ in range(local_world_size)] dist.gather(inputs, intra_output, dst=src_rank, group=intra_node_group) intra_output = [v.chunk(world_size, dim=0) for v in intra_output] intra_output = torch.cat(sum(zip(*intra_output), ())) # inter-node all-to-all if inter_node_group is not None: inter_output = torch.empty_like(intra_output) dist.all_to_all_single(inter_output, intra_output, group=inter_node_group) # layout transform inter_output = inter_output.chunk(num_group, dim=0) inter_output = [v.chunk(local_world_size, dim=0) for v in inter_output] intra_output = torch.cat(sum(zip(*inter_output), ())) # intra-node scatter intra_output = list(intra_output.chunk(local_world_size, dim=0)) dist.scatter(outputs, intra_output, src=src_rank, group=intra_node_group) else: dist.gather(inputs, dst=src_rank, group=intra_node_group) dist.scatter(outputs, src=src_rank, group=intra_node_group) return outputs @staticmethod def backward(ctx: Any, *grad_outputs) -> Tuple[Tensor, None, None]: return ( HierarchicalAllToAll.forward(None, grad_outputs[0], ctx.comm_grps, ctx.src_rank), None, None, ) class MoeDispatch(torch.autograd.Function): @staticmethod @custom_fwd def forward(ctx, tokens, mask, dest_idx, ec): s = tokens.size(0) h = tokens.size(1) dtype = tokens.dtype if MOE_KERNEL is None: load_moe() if tokens.dtype != torch.float32: tokens = tokens.to(torch.float32) expert_input = MOE_KERNEL.dispatch_forward(s, ec, h, tokens, mask, dest_idx) if expert_input.dtype != dtype: expert_input = expert_input.to(dtype) ctx.save_for_backward(mask, dest_idx) ctx.s = s ctx.h = h ctx.ec = ec ctx.dtype = dtype return expert_input @staticmethod @custom_bwd def backward(ctx, output_grad): mask, dest_idx = ctx.saved_tensors if output_grad.dtype != torch.float32: output_grad = output_grad.to(torch.float32) d_tokens = MOE_KERNEL.dispatch_backward(ctx.s, ctx.ec, ctx.h, output_grad, mask, dest_idx) if d_tokens.dtype != ctx.dtype: d_tokens = d_tokens.to(ctx.dtype) return d_tokens, None, None, None class MoeCombine(torch.autograd.Function): @staticmethod @custom_fwd def forward(ctx, expert_tokens, logits, mask, dest_idx, ec): assert logits.dtype == torch.float32 s = logits.size(0) e = logits.size(1) c = ec // e h = expert_tokens.size(-1) dtype = expert_tokens.dtype if expert_tokens.dtype != torch.float32: expert_tokens = expert_tokens.to(torch.float32) if MOE_KERNEL is None: load_moe() output = MOE_KERNEL.combine_forward(s, e, c, h, expert_tokens, logits, mask, dest_idx) if output.dtype != dtype: output = output.to(dtype) ctx.save_for_backward(expert_tokens, logits, mask, dest_idx) ctx.s = s ctx.e = e ctx.c = c ctx.h = h ctx.dtype = dtype return output @staticmethod @custom_bwd def backward(ctx, tokens_grad): expert_tokens, logits, mask, dest_idx = ctx.saved_tensors if tokens_grad.dtype != torch.float32: tokens_grad = tokens_grad.to(torch.float32) d_expert, d_logits = MOE_KERNEL.combine_backward( ctx.s, ctx.e, ctx.c, ctx.h, tokens_grad, expert_tokens, logits, mask, dest_idx ) if d_expert.dtype != ctx.dtype: d_expert = d_expert.to(ctx.dtype) return d_expert, d_logits, None, None, None def moe_cumsum(inputs: Tensor, use_kernel: bool = False): dim0 = inputs.size(0) flag = (dim0 <= 1024) or (dim0 <= 2048 and dim0 % 2 == 0) or (dim0 % 4 == 0) if flag and use_kernel: if MOE_KERNEL is None: load_moe() return MOE_KERNEL.cumsum_sub_one(inputs) else: return torch.cumsum(inputs, dim=0) - 1 class EPGradScalerIn(torch.autograd.Function): """ Scale the gradient back by the number of experts because the batch size increases in the moe stage """ @staticmethod def forward(ctx: Any, inputs: Tensor, ep_size: int) -> Tensor: ctx.ep_size = ep_size return inputs @staticmethod def backward(ctx: Any, *grad_outputs: Tensor) -> Tuple[Tensor, None]: assert len(grad_outputs) == 1 grad = grad_outputs[0] if ctx.ep_size != 1: grad.mul_(ctx.ep_size) return grad, None class EPGradScalerOut(torch.autograd.Function): """ Scale the gradient by the number of experts because the batch size increases in the moe stage """ @staticmethod def forward(ctx: Any, inputs: Tensor, ep_size: int) -> Tensor: ctx.ep_size = ep_size return inputs @staticmethod def backward(ctx: Any, *grad_outputs: Tensor) -> Tuple[Tensor, None]: assert len(grad_outputs) == 1 grad = grad_outputs[0] if ctx.ep_size != 1: grad.div_(ctx.ep_size) return grad, None class DPGradScalerIn(torch.autograd.Function): """ Scale the gradient back by the number of experts because the batch size increases in the moe stage """ @staticmethod def forward(ctx: Any, inputs: Tensor, moe_dp_size: int, activated_experts: int) -> Tensor: assert activated_experts != 0, f"shouldn't be called when no expert is activated" ctx.moe_dp_size = moe_dp_size ctx.activated_experts = activated_experts return inputs @staticmethod def backward(ctx: Any, *grad_outputs: Tensor) -> Tuple[Tensor, None, None]: assert len(grad_outputs) == 1 grad = grad_outputs[0] if ctx.moe_dp_size != ctx.activated_experts: grad.mul_(ctx.activated_experts / ctx.moe_dp_size) return grad, None, None class DPGradScalerOut(torch.autograd.Function): """ Scale the gradient by the number of experts because the batch size increases in the moe stage """ @staticmethod def forward(ctx: Any, inputs: Tensor, moe_dp_size: int, activated_experts: int) -> Tensor: assert activated_experts != 0, f"shouldn't be called when no expert is activated" ctx.moe_dp_size = moe_dp_size ctx.activated_experts = activated_experts return inputs @staticmethod def backward(ctx: Any, *grad_outputs: Tensor) -> Tuple[Tensor, None, None]: assert len(grad_outputs) == 1 grad = grad_outputs[0] if ctx.moe_dp_size != ctx.activated_experts: grad.mul_(ctx.moe_dp_size / ctx.activated_experts) return grad, None, None def _all_to_all( inputs: torch.Tensor, input_split_sizes: Optional[List[int]] = None, output_split_sizes: Optional[List[int]] = None, group=None, async_op: bool = False, fp8_communication: bool = False, ): """ Returns: outputs: Tensor handle: Optional[Work], if overlap is True """ outputs_shape = list(inputs.shape) if output_split_sizes is not None: outputs_shape[0] = sum(output_split_sizes) outputs = torch.empty(outputs_shape, dtype=inputs.dtype, device=inputs.device) inputs = inputs.contiguous() outputs = outputs.contiguous() if fp8_communication: handle = all_to_all_single_fp8( outputs, inputs, output_split_sizes, input_split_sizes, group=group, async_op=False ) else: handle = dist.all_to_all_single( outputs, inputs, output_split_sizes, input_split_sizes, group=group, async_op=async_op ) return outputs, handle class AllToAllUneven(torch.autograd.Function): @staticmethod def forward( ctx, inputs, input_split_sizes=None, output_split_sizes=None, group=None, overlap: bool = False, fp8_communication: bool = False, ): """ Returns: outputs: Tensor handle: Optional[Work], if overlap is True """ ctx.input_split_sizes = input_split_sizes ctx.output_split_sizes = output_split_sizes ctx.group = group return _all_to_all( inputs, input_split_sizes, output_split_sizes, group, overlap, fp8_communication=fp8_communication ) @staticmethod def backward(ctx: Any, *grad_outputs): return ( _all_to_all(grad_outputs[0], ctx.output_split_sizes, ctx.input_split_sizes, ctx.group, False)[0], None, None, None, None, None, ) def all_to_all_uneven( inputs: torch.Tensor, input_split_sizes: Optional[List[int]] = None, output_split_sizes: Optional[List[int]] = None, group=None, overlap: bool = False, fp8_communication: bool = False, ): return AllToAllUneven.apply(inputs, input_split_sizes, output_split_sizes, group, overlap, fp8_communication)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/moe/__init__.py
colossalai/moe/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/weight_grad_store.py
colossalai/pipeline/weight_grad_store.py
import queue class WeightGradStore: cache = [] weight_grad_queue = [queue.Queue(), queue.Queue()] @classmethod def put(cls, total_input, grad_output, weight, func): cls.cache.append((total_input, grad_output, weight, func)) @classmethod def flush(cls, chunk=0): cls.weight_grad_queue[chunk].put(cls.cache) cls.cache = [] @classmethod def pop(cls, chunk=0): if cls.weight_grad_queue[chunk].qsize() > 0: stored_grads = cls.weight_grad_queue[chunk].get() for total_input, grad_output, weight, func in stored_grads: if isinstance(weight, tuple): # In order to be hooked into Gemini's '__torch_function__', adding a view operation to weight and bias. # View will lead to weight ptr change # weight_cal & weight_origin in tuple, weight_cal use to cal dw, weight_origin use to update _, weight_origin = weight if weight_origin.grad is not None: func(total_input, grad_output, weight_origin.grad) # for first bwd; weight.grad is None, assign grad_weight to weight.grad else: grad_weight = func(total_input, grad_output) weight_origin.grad = grad_weight else: if weight.grad is not None: func(total_input, grad_output, weight.grad) # for first bwd; weight.grad is None, assign grad_weight to weight.grad else: grad_weight = func(total_input, grad_output) weight.grad = grad_weight else: raise Exception("Pop empty queue.")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/stage_manager.py
colossalai/pipeline/stage_manager.py
import contextlib from typing import Dict, List, Optional, Tuple, Union import numpy as np import torch.distributed as dist from torch.distributed import ProcessGroup from colossalai.cluster import ProcessGroupMesh class PipelineStageManager: """PipelineStageManager is a helper class to manage pipeline stages. Args: pg_mesh (ProcessGroupMesh): Process group mesh. pipeline_axis (int): The axis along which the pipeline is constructed. is_virtual (bool): Whether to use circle p2p communication, it will make the first and last stage communicate with each other. Attributes: num_stages (int): Number of stages in the pipeline. stage (int): The current stage. """ def __init__( self, pg_mesh: ProcessGroupMesh, pipeline_axis: int, enable_interleave: bool = False, use_zbv: bool = False, num_model_chunks: int = 1, num_layers_per_stage: Optional[List[int]] = None, ) -> None: assert enable_interleave or num_model_chunks == 1, "num_model_chunks must be 1 when enable_interleave is False" self.pg_mesh = pg_mesh self.pipeline_axis = pipeline_axis self.prev_rank: Optional[Tuple[int, ...]] = None self.next_rank: Optional[Tuple[int, ...]] = None self.p2p_groups: Dict[Tuple[int, ...], ProcessGroup] = {} if num_layers_per_stage is not None: assert len(num_layers_per_stage) == self.num_stages self.num_layers_per_stage = num_layers_per_stage # init prev and next coord coord = self.pg_mesh.coordinate() # the prev rank of rank0 is the last rank prev_coord = coord[: self.pipeline_axis] + (coord[self.pipeline_axis] - 1,) + coord[self.pipeline_axis + 1 :] self.prev_rank = self.pg_mesh.ravel(prev_coord, self.pg_mesh.shape, mode="wrap") # the next rank of the last rank is rank0 next_coord = coord[: self.pipeline_axis] + (coord[self.pipeline_axis] + 1,) + coord[self.pipeline_axis + 1 :] self.next_rank = self.pg_mesh.ravel(next_coord, self.pg_mesh.shape, mode="wrap") self.is_interleave = enable_interleave self.use_zbv = use_zbv # for interleaved pipeline parallel, each device is responsible for multiple chunk of layers self.num_model_chunks: int = num_model_chunks # for shardformer, hold stage indices of model self.stage_indices: List[Tuple[int, int]] # for shardformer, hold model chunk id self.model_chunk_id: Optional[int] = None self.p2p_group = self.pg_mesh.get_group_along_axis(self.pipeline_axis) def get_stage_index( self, layers_per_stage: List[int], stage: Optional[int] = None, num_model_chunks: Optional[int] = None, num_stages: Optional[int] = None, ) -> Union[Tuple[int, int], List[Tuple[int, int]]]: """ Get the start index and end index of layers for each stage. Args: layers_per_stage (List[int]): number of layers for each stage stage (int): the stage index num_stages (int): number of stages num_model_chunks (int): number of model chunks Returns: - Tuple[int, int]: the start index and end index of this stage - List[Tuple[int, int]]: the start index and end index of this stage for each model chunk """ stage = self.stage if stage is None else stage num_model_chunks = self.num_model_chunks if num_model_chunks is None else num_model_chunks num_stages = self.num_stages if num_stages is None else num_stages num_layers_per_stage_accumulated = np.insert(np.cumsum(layers_per_stage), 0, 0) stage_indices = [] if self.use_zbv: stage_indices.append([num_layers_per_stage_accumulated[stage], num_layers_per_stage_accumulated[stage + 1]]) stage_indices.append( [ num_layers_per_stage_accumulated[2 * num_stages - stage - 1], num_layers_per_stage_accumulated[2 * num_stages - stage], ] ) return stage_indices for model_chunk in range(num_model_chunks): start_idx = num_layers_per_stage_accumulated[stage + model_chunk * num_stages] end_idx = num_layers_per_stage_accumulated[stage + model_chunk * num_stages + 1] stage_indices.append([start_idx, end_idx]) return stage_indices[0] if num_model_chunks == 1 else stage_indices def is_first_stage(self, ignore_chunk: bool = False) -> bool: """Is the current stage the first stage. NOTE: 1. if using interleaved pipeline parallel, the first stage is the first chunk of the first device. 2. invoke is_first_stage() with ignore_chunk=True is equivalent to invoke is_first_device() Returns: bool: Whether the current stage is the first stage. """ assert isinstance(ignore_chunk, bool) assert not self.is_interleave or (ignore_chunk or self.model_chunk_id is not None) if not self.is_interleave or ignore_chunk: return self.stage == 0 else: return self.stage == 0 and self.model_chunk_id == 0 def is_last_stage(self, ignore_chunk: bool = False) -> bool: """Is the current stage the last stage. NOTE: 1. if using interleaved pipeline parallel, the last stage is the last chunk of the last device. 2. invoke is_last_stage() with ignore_chunk=True is equivalent to invoke is_last_device() Returns: bool: Whether the current stage is the last stage. """ assert isinstance(ignore_chunk, bool) assert not self.is_interleave or (ignore_chunk or self.model_chunk_id is not None) if not self.is_interleave or ignore_chunk: return self.stage == self.num_stages - 1 else: # use zero bubble pipeline if self.use_zbv: return self.stage == 0 and self.model_chunk_id == self.num_model_chunks - 1 else: return self.stage == self.num_stages - 1 and self.model_chunk_id == self.num_model_chunks - 1 @property def num_stages(self) -> int: """Number of stages in the pipeline. Returns: int: Number of stages in the pipeline. """ return self.pg_mesh.size(self.pipeline_axis) @property def stage(self) -> int: """Current stage. Returns: int: Current stage. """ return self.pg_mesh.coordinate(self.pipeline_axis) def get_rank(self) -> int: """Get the rank of the current process. Returns: int: Rank of the current process. """ return dist.get_rank() def get_prev_rank(self) -> int: """Get the rank of the previous stage. Returns: int: Rank of the previous stage. """ return self.prev_rank def get_next_rank(self) -> int: """Get the rank of the next stage. Returns: int: Rank of the next stage. """ return self.next_rank def get_p2p_process_group(self) -> ProcessGroup: """Get the p2p process group between two ranks. The order of the two ranks does not matter. Returns: ProcessGroup: P2P process group between the two ranks. """ return self.p2p_group def init_process_group_by_stages(self, stages: List[int]) -> ProcessGroup: """Get the process group of the given stages. Args: stages (List[int]): List of stages. Returns: ProcessGroup: Process group of the given stages. """ return self.pg_mesh.get_group_along_axis(self.pipeline_axis, stages) @contextlib.contextmanager def switch_model_chunk_id(self, model_chunk_id: int): old_model_chunk_id = self.model_chunk_id self.model_chunk_id = model_chunk_id yield self.model_chunk_id = old_model_chunk_id def distribute_layers( self, num_layers: int, num_stages: Optional[int] = None, num_model_chunks: Optional[int] = None ) -> List[int]: if self.num_layers_per_stage is not None: assert sum(self.num_layers_per_stage) == num_layers return self.num_layers_per_stage num_stages = self.num_stages if num_stages is None else num_stages num_model_chunks = self.num_model_chunks if num_model_chunks is None else num_model_chunks quotient = num_layers // (num_stages * num_model_chunks) remainder = num_layers % (num_stages * num_model_chunks) # calculate the num_layers per stage layers_per_stage = [quotient] * num_stages * num_model_chunks # deal with the rest layers if remainder > 0: start_position = (num_stages * num_model_chunks) // 2 - remainder // 2 for i in range(start_position, start_position + remainder): layers_per_stage[i] += 1 return layers_per_stage
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/__init__.py
colossalai/pipeline/__init__.py
from .p2p import PipelineP2PCommunication from .schedule import InterleavedSchedule, OneForwardOneBackwardSchedule, PipelineSchedule, ZeroBubbleVPipeScheduler from .stage_manager import PipelineStageManager __all__ = [ "PipelineSchedule", "OneForwardOneBackwardSchedule", "InterleavedSchedule", "ZeroBubbleVPipeScheduler", "PipelineP2PCommunication", "PipelineStageManager", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/p2p.py
colossalai/pipeline/p2p.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- import io import pickle import re from collections import namedtuple from typing import Any, Callable, List, Optional, Tuple, Union import torch import torch.distributed as dist from packaging.version import Version from torch.distributed import ProcessGroup from torch.distributed import distributed_c10d as c10d from torch.utils._pytree import tree_flatten, tree_unflatten from colossalai.accelerator import get_accelerator from .stage_manager import PipelineStageManager def _cuda_safe_tensor_to_object(tensor: torch.Tensor, tensor_size: torch.Size) -> Any: """transform tensor to object with unpickle. Info of the device in bytes stream will be modified into current device before unpickling Args: tensor (:class:`torch.tensor`): tensor to be unpickled tensor_size (:class:`torch.Size`): Size of the real info in bytes Returns: Any: object after unpickled """ buf = tensor.numpy().tobytes()[:tensor_size] if b"cuda" in buf: buf_array = bytearray(buf) device_index = get_accelerator().current_device() # There might be more than one output tensors during forward for cuda_str in re.finditer(b"cuda", buf_array): pos = cuda_str.start() buf_array[pos + 5] = 48 + device_index buf = bytes(buf_array) io_bytes = io.BytesIO(buf) byte_pickler = pickle.Unpickler(io_bytes) unpickle = byte_pickler.load() return unpickle def check_for_nccl_backend(group): pg = group or c10d._get_default_group() # Gate PG wrapper check on Gloo availability. if c10d._GLOO_AVAILABLE: # It is not expected for PG to be wrapped many times, but support it just # in case while isinstance(pg, c10d._ProcessGroupWrapper): pg = pg.wrapped_pg return c10d.is_nccl_available() and pg.name() == c10d.Backend.NCCL # NOTE: FIXME: NPU DOES NOT support isend nor irecv, so broadcast is kept for future use def _broadcast_object_list( object_list: List[Any], src: int, group: ProcessGroup, device: Optional[Union[torch.device, str, int]] = None ): """This is a modified version of the broadcast_object_list in torch.distribution The only difference is that object will be move to correct device after unpickled. If local_rank = src, then object list will be sent to rank src. Otherwise, object list will be updated with data sent from rank src. Args: object_list (List[Any]): list of object to broadcast src (int): source rank to broadcast dst (int): dst rank to broadcast device (:class:`torch.device`): device to do broadcast. current device in default """ if c10d._rank_not_in_group(group): c10d._warn_not_in_group("broadcast_object_list") return is_nccl_backend = _check_for_nccl_backend(group) current_device = None if device is not None: if is_nccl_backend and device.type != "cuda": raise ValueError("device type must be cuda for nccl backend") current_device = device else: current_device = torch.device("cpu") if is_nccl_backend: current_device = torch.device("cuda", get_accelerator().current_device()) my_rank = dist.get_rank() # Serialize object_list elements to tensors on src rank. if my_rank == src: if Version(torch.__version__) >= Version("2.3.0"): tensor_list, size_list = zip( *[c10d._object_to_tensor(obj, device=current_device, group=group) for obj in object_list] ) elif Version(torch.__version__) >= Version("1.13.0"): tensor_list, size_list = zip(*[c10d._object_to_tensor(obj, device=current_device) for obj in object_list]) else: tensor_list, size_list = zip(*[c10d._object_to_tensor(obj) for obj in object_list]) object_sizes_tensor = torch.cat(size_list) else: object_sizes_tensor = torch.empty(len(object_list), dtype=torch.long) if is_nccl_backend: object_sizes_tensor = object_sizes_tensor.to(current_device) # Broadcast object sizes c10d.broadcast(object_sizes_tensor, src=src, group=group, async_op=False) # Concatenate and broadcast serialized object tensors if my_rank == src: object_tensor = torch.cat(tensor_list) else: object_tensor = torch.empty( # type: ignore[call-overload] torch.sum(object_sizes_tensor).item(), # type: ignore[arg-type] dtype=torch.uint8, ) if is_nccl_backend: object_tensor = object_tensor.to(current_device) c10d.broadcast(object_tensor, src=src, group=group, async_op=False) # Deserialize objects using their stored sizes. offset = 0 if my_rank != src: for i, obj_size in enumerate(object_sizes_tensor): obj_view = object_tensor[offset : offset + obj_size] obj_view = obj_view.type(torch.uint8) if obj_view.device != torch.device("cpu"): obj_view = obj_view.cpu() offset += obj_size # unpickle unpickle_object = _cuda_safe_tensor_to_object(obj_view, obj_size) # unconsistence in device if ( isinstance(unpickle_object, torch.Tensor) and unpickle_object.device.index != get_accelerator().current_device() ): unpickle_object = unpickle_object.to(get_accelerator().current_device()) object_list[i] = unpickle_object def _check_for_nccl_hccl_backend(group): pg = group or c10d._get_default_group() # Gate PG wrapper check on Gloo availability. if c10d._GLOO_AVAILABLE: # It is not expected for PG to be wrapped many times, but support it just in case while isinstance(pg, c10d._ProcessGroupWrapper): pg = pg.wrapped_pg return (c10d.is_nccl_available() or torch.distributed.is_hccl_available()) and ( pg.name() == c10d.Backend.NCCL or pg.name() == c10d.Backend.HCCL ) def _check_device(group): is_nccl_backend = _check_for_nccl_hccl_backend(group) current_device = torch.device("cpu") if is_nccl_backend: current_device = torch.device(get_accelerator().current_device()) return current_device, is_nccl_backend TensorMetadata = namedtuple("TensorMetadata", ["shape", "dtype", "requires_grad"]) P2PMetadata = namedtuple("P2PMetadata", ["tree_spec", "tensor_metadata", "non_tensor_obj_idx", "non_tensor_objs"]) def create_send_metadata( object: Any, strict: bool = True, return_tensor: bool = False ) -> Union[P2PMetadata, Tuple[P2PMetadata, List[torch.Tensor]]]: """ Args: object (Any): object needed to be sent strict (bool, optional): whether to check if the object is supported for fast send return_tensor (bool, optional): whether to return tensor objects """ objs, tree_spec = tree_flatten(object) tensor_metadata, tensor_objs = [], [] non_tensor_obj_idx, non_tensor_objs = [], [] for idx, obj in enumerate(objs): if isinstance(obj, torch.Tensor): tensor_objs.append(obj) tensor_metadata.append(TensorMetadata(obj.shape, obj.dtype, obj.requires_grad)) else: non_tensor_obj_idx.append(idx) non_tensor_objs.append(obj) assert not strict or len(non_tensor_objs) == 0, "Only support tensor for fast send" metadata = P2PMetadata(tree_spec, tensor_metadata, non_tensor_obj_idx, non_tensor_objs) return metadata if not return_tensor else (metadata, tensor_objs) def _filling_ops_queue( obj: Union[torch.Tensor, List[torch.Tensor]], comm_op: Callable, comm_rank: int, ops_queue: List, group: ProcessGroup, ): if isinstance(obj, torch.Tensor): obj = obj.contiguous() op_to_add = dist.P2POp(comm_op, obj, comm_rank, group) ops_queue.append(op_to_add) else: for tensor_to_comm in obj: assert isinstance(tensor_to_comm, torch.Tensor) _filling_ops_queue(tensor_to_comm, comm_op, comm_rank, ops_queue, group) def _create_recv_buffer(tensor_metadata: List[TensorMetadata], current_device) -> List[torch.Tensor]: buffer_recv = [] for metadata in tensor_metadata: tensor_recv = torch.empty( metadata.shape, requires_grad=metadata.requires_grad, device=current_device, dtype=metadata.dtype ) buffer_recv.append(tensor_recv) return buffer_recv def _batch_send_recv_tensor( send_tensor_list: Optional[List[torch.Tensor]], recv_tensor_metadata: Optional[List[TensorMetadata]], send_dst: Optional[int], recv_src: Optional[int], send_group: Optional[ProcessGroup], recv_group: Optional[ProcessGroup], current_device: Any, overlap_p2p: bool = True, send_first: bool = True, ) -> Optional[Union[torch.Tensor, List[torch.Tensor]]]: buffer_recv = None if recv_tensor_metadata is not None: buffer_recv = _create_recv_buffer(recv_tensor_metadata, current_device) ops = [] is_send = send_dst is not None and send_tensor_list is not None is_recv = recv_src is not None and buffer_recv is not None if send_first: if is_send: assert send_group is not None _filling_ops_queue(send_tensor_list, dist.isend, send_dst, ops, send_group) if is_recv: assert recv_group is not None _filling_ops_queue(buffer_recv, dist.irecv, recv_src, ops, recv_group) else: if is_recv: assert recv_group is not None _filling_ops_queue(buffer_recv, dist.irecv, recv_src, ops, recv_group) if is_send: assert send_group is not None _filling_ops_queue(send_tensor_list, dist.isend, send_dst, ops, send_group) if len(ops) > 0: reqs = dist.batch_isend_irecv(ops) if not overlap_p2p: for req in reqs: req.wait() return buffer_recv, [] else: return buffer_recv, reqs return None, [] def _send_recv_serialization_object( object: Optional[P2PMetadata], send_dst: Optional[int], recv_src: Optional[int], send_group: Optional[ProcessGroup], recv_group: Optional[ProcessGroup], current_device: Any, is_nccl_backend: bool, send_first: bool = True, ) -> Optional[P2PMetadata]: ops = [] send_object_tensor = None send_object_size_tensor = None if object is not None and send_dst is not None: if Version(torch.__version__) >= Version("2.3.0"): send_object_tensor, send_object_size_tensor = c10d._object_to_tensor( object, device=current_device, group=send_group ) elif Version(torch.__version__) >= Version("1.13.0"): send_object_tensor, send_object_size_tensor = c10d._object_to_tensor(object, device=current_device) else: send_object_tensor, send_object_size_tensor = c10d._object_to_tensor(object) if is_nccl_backend: send_object_size_tensor = send_object_size_tensor.to(current_device) send_object_tensor = send_object_tensor.to(current_device) recv_object_size_tensor = None if recv_src is not None: recv_object_size_tensor = torch.empty(1, dtype=torch.long) if is_nccl_backend: recv_object_size_tensor = recv_object_size_tensor.to(current_device) if send_first: if send_object_size_tensor is not None: _filling_ops_queue(send_object_size_tensor, dist.isend, send_dst, ops, send_group) if recv_src is not None: _filling_ops_queue(recv_object_size_tensor, dist.irecv, recv_src, ops, recv_group) else: if recv_src is not None: _filling_ops_queue(recv_object_size_tensor, dist.irecv, recv_src, ops, recv_group) if send_object_size_tensor is not None: _filling_ops_queue(send_object_size_tensor, dist.isend, send_dst, ops, send_group) if len(ops) > 0: reqs = dist.batch_isend_irecv(ops) for req in reqs: req.wait() # This blocks the compute stream in torch ops = [] is_send = send_dst is not None and send_object_tensor is not None is_recv = recv_src is not None and recv_object_size_tensor is not None recv_object_tensor = None if is_recv: recv_object_tensor = torch.empty(recv_object_size_tensor.item(), dtype=torch.uint8) if is_nccl_backend: recv_object_tensor = recv_object_tensor.to(current_device) if send_first: if is_send: _filling_ops_queue(send_object_tensor, dist.isend, send_dst, ops, send_group) if is_recv: _filling_ops_queue(recv_object_tensor, dist.irecv, recv_src, ops, recv_group) else: if is_recv: _filling_ops_queue(recv_object_tensor, dist.irecv, recv_src, ops, recv_group) if is_send: _filling_ops_queue(send_object_tensor, dist.isend, send_dst, ops, send_group) if len(ops) > 0: reqs = dist.batch_isend_irecv(ops) for req in reqs: req.wait() if recv_object_tensor is not None and recv_object_size_tensor is not None: recv_object_tensor = recv_object_tensor.type(torch.uint8) if recv_object_tensor.device != torch.device("cpu"): recv_object_tensor = recv_object_tensor.cpu() unpickle_object = _cuda_safe_tensor_to_object(recv_object_tensor, recv_object_size_tensor.item()) if ( isinstance(unpickle_object, torch.Tensor) and unpickle_object.device.index != get_accelerator().current_device() ): unpickle_object = unpickle_object.to(get_accelerator().current_device()) return unpickle_object def _communicate( object: Any, send_dst: Optional[int], recv_src: Optional[int], overlap_p2p: bool, send_group: Optional[ProcessGroup] = None, recv_group: Optional[ProcessGroup] = None, send_metadata: bool = True, metadata_recv: Optional[P2PMetadata] = None, send_first: Optional[bool] = None, ) -> Any: """ Send and receive object from send_dst and recv_src respectively Args: object (Any): object needed to be sent send_dst (int): rank of the destination recv_src (int): rank of the source overlap_p2p (bool): whether to overlap p2p communication with computation send_group (ProcessGroup, optional): process group of sender recv_group (ProcessGroup, optional): process group of receiver send_metadata (bool, optional): whether to send metadata metadata_recv (P2PMetadata, optional): metadata of the object to be received """ assert send_dst is not None or recv_src is not None, "send_dst and recv_src cannot be both None" assert send_dst is None or send_group is not None, "send_group must be specified when send_dst is not None" assert recv_src is None or recv_group is not None, "recv_group must be specified when recv_src is not None" assert ( metadata_recv is None or len(metadata_recv.non_tensor_obj_idx) == 0 ), "metadata_recv should not contain non-tensor objects" metadata_send, tensor_objs = None, None if object is not None: # NOTE: if object contains non-tensor objects, we have to send metadata metadata_send, tensor_objs = create_send_metadata(object, strict=False, return_tensor=True) send_metadata = send_metadata or len(metadata_send.non_tensor_obj_idx) > 0 else: send_metadata = False assert not c10d._rank_not_in_group(send_group) and not c10d._rank_not_in_group(recv_group) current_send_device, is_send_nccl_backend = _check_device(send_group) current_recv_device, is_recv_nccl_backend = _check_device(recv_group) is_nccl_backend = is_send_nccl_backend and is_recv_nccl_backend assert current_send_device == current_recv_device current_device = current_send_device if (send_dst is not None and send_metadata) or (recv_src is not None and metadata_recv is None): # Send and receive metadata _metadata_recv = _send_recv_serialization_object( object=metadata_send, send_dst=send_dst if send_metadata else None, recv_src=recv_src if metadata_recv is None else None, send_group=send_group if send_metadata else None, recv_group=recv_group if metadata_recv is None else None, current_device=current_device, is_nccl_backend=is_nccl_backend, send_first=send_first if send_first != None else True, ) assert ( metadata_recv is None or _metadata_recv is None ), "You shouldn't receive metadata when using the cached metadata" metadata_recv = _metadata_recv if metadata_recv is None else metadata_recv # Send and receive data recv_tensor_metadata = None if metadata_recv is None else metadata_recv.tensor_metadata recv_tensor_objs, wait_handles = _batch_send_recv_tensor( tensor_objs, recv_tensor_metadata, send_dst, recv_src, send_group, recv_group, current_device, overlap_p2p=overlap_p2p, send_first=send_first if send_first != None else True, ) if metadata_recv is not None: assert isinstance(metadata_recv, P2PMetadata) tree_spec = metadata_recv.tree_spec non_tensor_obj_idx = metadata_recv.non_tensor_obj_idx non_tensor_objs = metadata_recv.non_tensor_objs if recv_tensor_objs is None: recv_tensor_objs = [] for idx in non_tensor_obj_idx: recv_tensor_objs.insert(idx, non_tensor_objs.pop(0)) recv_object = tree_unflatten(recv_tensor_objs, tree_spec) return recv_object, wait_handles return None, wait_handles def _p2p_comm( tensor_send_next: torch.Tensor, recv_prev: bool, peer: int, group: ProcessGroup, comm_dtype: torch.dtype = torch.float16, ): """ Send and recv tensor using P2P communication, used when pipeline size is 2 to solve the race communication. Args: tensor_send_next (torch.Tensor): tensor to be sent to next stage recv_prev (bool): whether to receive tensor from previous stage peer (int): rank of the peer group (ProcessGroup): process group comm_dtype (torch.dtype): dtype of the tensor to be sent Returns: torch.Tensor: tensor received from previous stage """ # send and recv shape send_next_shape = None recv_prev_shape = None if tensor_send_next is not None: send_next_shape = torch.tensor( tensor_send_next.size(), device=get_accelerator().current_device(), dtype=torch.int64 ) if recv_prev: recv_prev_shape = torch.empty((3), device=get_accelerator().current_device(), dtype=torch.int64) ops = [] if send_next_shape is not None: send_next_op = dist.P2POp(dist.isend, send_next_shape, peer=peer, group=group) ops.append(send_next_op) if recv_prev_shape is not None: recv_prev_op = dist.P2POp( dist.irecv, recv_prev_shape, peer=peer, group=group, ) ops.append(recv_prev_op) if len(ops) > 0: reqs = dist.batch_isend_irecv(ops) for req in reqs: req.wait() if recv_prev_shape is not None: recv_prev_shape = recv_prev_shape.tolist() # send and recv data tensor_recv_prev = None if recv_prev: tensor_recv_prev = torch.empty(recv_prev_shape, device=get_accelerator().current_device(), dtype=comm_dtype) ops = [] if tensor_send_next is not None: send_next_op = dist.P2POp( dist.isend, tensor_send_next, peer=peer, group=group, ) ops.append(send_next_op) if tensor_recv_prev is not None: recv_prev_op = dist.P2POp( dist.irecv, tensor_recv_prev, peer=peer, group=group, ) ops.append(recv_prev_op) if len(ops) > 0: reqs = dist.batch_isend_irecv(ops) for req in reqs: req.wait() return tensor_recv_prev class PipelineP2PCommunication: def __init__(self, stage_manager: PipelineStageManager, overlap_p2p: bool = True) -> None: self.stage_manager = stage_manager self.overlap_p2p = overlap_p2p def recv_forward( self, prev_rank: Optional[int] = None, metadata_recv: Optional[P2PMetadata] = None ) -> Tuple[Any, List]: """Copy the forward output from the previous stage in pipeline as the input tensor of this stage. Args: prev_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ if prev_rank is None: prev_rank = self.stage_manager.get_prev_rank() input_tensor, wait_handles = _communicate( object=None, recv_src=prev_rank, send_dst=None, recv_group=self.stage_manager.get_p2p_process_group(), metadata_recv=metadata_recv, overlap_p2p=self.overlap_p2p, ) return input_tensor, wait_handles def recv_backward( self, next_rank: Optional[int] = None, metadata_recv: Optional[P2PMetadata] = None ) -> Tuple[Any, List]: """Copy the gradient tensor from the next stage in pipeline as the input gradient of this stage. Args: next_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ if next_rank is None: next_rank = self.stage_manager.get_next_rank() output_tensor_grad, wait_handles = _communicate( object=None, recv_src=next_rank, send_dst=None, recv_group=self.stage_manager.get_p2p_process_group(), metadata_recv=metadata_recv, overlap_p2p=self.overlap_p2p, ) return output_tensor_grad, wait_handles def send_forward(self, output_object: Any, next_rank: Optional[int] = None, send_metadata: bool = True) -> List: """Sends the input tensor to the next stage in pipeline. Args: output_object (Any): Object to be sent. next_rank (int, optional): The rank of the recipient of the tensor. Returns: List: List of handles for the communication requests, if overlap is enabled. """ if next_rank is None: next_rank = self.stage_manager.get_next_rank() _, handles = _communicate( output_object, recv_src=None, send_dst=next_rank, send_group=self.stage_manager.get_p2p_process_group(), send_metadata=send_metadata, overlap_p2p=self.overlap_p2p, ) return handles def send_backward(self, input_object: Any, prev_rank: Optional[int] = None, send_metadata: bool = True) -> List: """Sends the gradient tensor to the previous stage in pipeline. Args: input_object (Any): Object to be sent. prev_rank (int, optional): The rank of the recipient of the tensor Returns: List: List of handles for the communication requests, if overlap is enabled. """ if prev_rank is None: prev_rank = self.stage_manager.get_prev_rank() _, handles = _communicate( input_object, recv_src=None, send_dst=prev_rank, send_group=self.stage_manager.get_p2p_process_group(), send_metadata=send_metadata, overlap_p2p=self.overlap_p2p, ) return handles def send_forward_recv_forward( self, output_object: Any, is_send: bool, is_recv: bool, send_first: bool, send_metadata: bool = True, metadata_recv: Optional[P2PMetadata] = None, ) -> Tuple[Any, List]: """Sends the input tensor to the next pipeline stage and copy the output tensor from the next pipeline stage Args: output_object (Any): Object to be sent. is_send (bool): Whether to send the input tensor to the next pipeline stage. is_recv (bool): Whether to copy the output tensor from the next pipeline stage. send_first (bool): Whether to send before receive. send_metadata (bool, optional): Whether to send metadata. metadata_recv (P2PMetadata, optional): The cached metadata(size, type) of the object to be received. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ next_rank = self.stage_manager.get_next_rank() if is_send else None prev_rank = self.stage_manager.get_prev_rank() if is_recv else None group = self.stage_manager.get_p2p_process_group() return _communicate( output_object, send_dst=next_rank, recv_src=prev_rank, send_group=group if is_send else None, recv_group=group if is_recv else None, send_metadata=send_metadata if is_send else False, metadata_recv=metadata_recv if is_recv else None, send_first=send_first, overlap_p2p=self.overlap_p2p, ) def send_backward_recv_backward( self, input_object: Any, is_send: bool, is_recv: bool, send_first: bool, send_metadata: bool = True, metadata_recv: Optional[P2PMetadata] = None, ) -> Tuple[Any, List]: """Sends the gradient tensor to the previous pipeline stage and copy the gradient tensor from the previous pipeline stage Args: input_object (Any): Object to be sent. is_send (bool): Whether to send the gradient tensor to the previous pipeline stage. is_recv (bool): Whether to copy the gradient tensor from the previous pipeline stage. send_first (bool): Whether to send before receive. send_metadata (bool, optional): Whether to send metadata. metadata_recv (P2PMetadata, optional): The cached metadata(size, type) of the object to be received. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ prev_rank = self.stage_manager.get_prev_rank() if is_send else None next_rank = self.stage_manager.get_next_rank() if is_recv else None group = self.stage_manager.get_p2p_process_group() return _communicate( input_object, send_dst=prev_rank, recv_src=next_rank, send_group=group if is_send else None, recv_group=group if is_recv else None, send_metadata=send_metadata if is_send else False, metadata_recv=metadata_recv if is_recv else None, send_first=send_first, overlap_p2p=self.overlap_p2p, ) def send_forward_recv_backward( self, input_object: Any, send_metadata: bool = True, metadata_recv: Optional[P2PMetadata] = None, send_first: Optional[bool] = None, ) -> Tuple[Any, List]: """Sends the gradient tensor to and copy the gradient tensor from the next pipeline stage Args: input_object (Any): Object to be sent. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ next_rank = self.stage_manager.get_next_rank() group = self.stage_manager.get_p2p_process_group() return _communicate( input_object, next_rank, next_rank, send_group=group, recv_group=group, send_metadata=send_metadata, metadata_recv=metadata_recv, send_first=send_first, overlap_p2p=False, ) def send_backward_recv_forward( self, input_object: Any, send_metadata: bool = True, metadata_recv: Optional[P2PMetadata] = None, send_first: Optional[bool] = None, ) -> Tuple[Any, List]: """Sends the gradient tensor to and copy the gradient tensor from the previous stage in pipeline Args: input_object (Any): Object to be sent. Returns: Any: The input tensor or input tensor list. List: List of handles for the communication requests, if overlap is enabled. """ prev_rank = self.stage_manager.get_prev_rank() group = self.stage_manager.get_p2p_process_group() return _communicate( input_object, prev_rank, prev_rank, send_group=group, recv_group=group, send_metadata=send_metadata, metadata_recv=metadata_recv, send_first=send_first, overlap_p2p=False, ) def p2p_communicate( self, output_object: Any, recv_pre: bool, next_rank: Optional[int] = None, comm_dtype: torch.dtype = torch.float16, ) -> Any: """ Sends the input tensor to the next stage in pipeline, using `P2Pop` in torch. Args: output_object (Any): Object to be sent. next_rank (int, optional): The rank of the recipient of the tensor. """ if next_rank is None: next_rank = self.stage_manager.get_next_rank() recv_tensor = _p2p_comm( output_object, recv_pre, next_rank, self.stage_manager.get_p2p_process_group(), comm_dtype, ) return recv_tensor
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/one_f_one_b.py
colossalai/pipeline/schedule/one_f_one_b.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Union import torch from torch.nn import Module from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.interface import ModelWrapper, OptimizerWrapper from colossalai.pipeline.p2p import PipelineP2PCommunication, create_send_metadata from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.quantization.fp8 import cast_from_fp8_pipeline, cast_to_fp8_pipeline from colossalai.utils import get_current_device from ._utils import ( detach, get_batch_size, get_micro_batch, merge_batch, model_forward, retain_grad, to_device, tree_map_hf, ) from .base import PipelineSchedule class OneForwardOneBackwardSchedule(PipelineSchedule): def __init__( self, stage_manager: PipelineStageManager, num_microbatches: Optional[int] = None, microbatch_size: Optional[int] = None, enable_metadata_cache: bool = True, fp8_communication: bool = False, ) -> None: """1F1B pipeline schedule. Args: stage_manager (PipelineStageManager): Pipeline stage manager num_microbatches (Optional[int], optional): The number of microbatches. If not provided, it will be derived from microbatch size. Defaults to None. microbatch_size (Optional[int], optional): Microbatch size. If num_microbatches is provided, this will be ignored. Defaults to None. """ super().__init__(stage_manager) assert ( num_microbatches is not None or microbatch_size is not None ), "Either num_microbatches or microbatch_size should be provided" self.comm = PipelineP2PCommunication(stage_manager, overlap_p2p=False) self.num_microbatches = num_microbatches self.microbatch_size = microbatch_size self.batch: Optional[Any] = None self.batch_size: Optional[int] = None self.last_batch_size: Optional[int] = None self.microbatch_offset: Optional[int] = None # P2PMeta cache self.enable_metadata_cache = enable_metadata_cache self.send_tensor_metadata = True self.send_grad_metadata = True self.tensor_metadata_recv = None self.grad_metadata_recv = None self.fp8_communication = fp8_communication def load_batch(self, data_iter: Iterable, device: Optional[torch.device] = None) -> None: """Load a batch from data iterator. Args: data_iter (Iterable): Data iterator. device (Optional[torch.device], optional): Target device. Defaults to None. """ batch = next(data_iter) if device is not None: batch = tree_map(partial(to_device, device=device), batch) self.microbatch_offset = 0 self.batch = batch self.batch_size = get_batch_size(batch) if self.microbatch_size is None: assert self.batch_size % self.num_microbatches == 0, "Batch size should divided by # microbatches" self.microbatch_size = self.batch_size // self.num_microbatches if self.num_microbatches is None: assert self.batch_size % self.microbatch_size == 0, "Batch size should divided by the microbatch size" self.num_microbatches = self.batch_size // self.microbatch_size if not self.forward_only: assert self.last_batch_size is None or self.last_batch_size == self.batch_size assert self.batch_size == self.microbatch_size * self.num_microbatches assert ( self.num_microbatches >= self.stage_manager.num_stages ), "Number of microbatch should be larger than number of stages" if self.forward_only: self.num_microbatches = (self.batch_size - 1) // self.microbatch_size + 1 # NOTE: disable metadata cache when batch size changes (not valid anymore) if self.batch_size != self.last_batch_size: self.enable_metadata_cache = False self.send_tensor_metadata = True self.send_grad_metadata = True self.tensor_metadata_recv = None self.grad_metadata_recv = None self.last_batch_size = self.batch_size def load_micro_batch(self) -> Any: """Load a micro batch from the current batch. Returns: Any: Micro batch. """ assert self.microbatch_offset <= self.batch_size, "Microbatches exhausted" micro_batch = get_micro_batch(self.batch, self.microbatch_offset, self.microbatch_size) self.microbatch_offset += self.microbatch_size return tree_map(partial(to_device, device=get_accelerator().get_current_device()), micro_batch) def recv_forward(self, prev_rank: int = None) -> Any: """Copy the forward output from the previous stage in pipeline as the input tensor of this stage. For 1F1B. Args: prev_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input tensor or input tensor list. """ if not self.stage_manager.is_first_stage(): input_tensor, _ = self.comm.recv_forward(prev_rank, metadata_recv=self.tensor_metadata_recv) if self.enable_metadata_cache and self.tensor_metadata_recv is None: self.tensor_metadata_recv = create_send_metadata(input_tensor) if self.fp8_communication: cast_from_fp8_pipeline(input_tensor) return input_tensor def recv_backward(self, next_rank: int = None) -> Any: """Copy the gradient tensor from the next stage in pipeline as the input gradient of this stage. For 1F1B. Args: next_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input gradient tensor or gradient tensor list. """ if not self.stage_manager.is_last_stage(): output_tensor_grad, _ = self.comm.recv_backward(next_rank, metadata_recv=self.grad_metadata_recv) if self.fp8_communication: cast_from_fp8_pipeline(output_tensor_grad) if self.enable_metadata_cache and self.grad_metadata_recv is None: self.grad_metadata_recv = create_send_metadata(output_tensor_grad) return output_tensor_grad def send_forward(self, output_tensor: Any, next_rank: int = None) -> None: """Sends the input tensor to the next stage in pipeline. For 1F1B. Args: output_object (Any): Object to be sent. next_rank (int, optional): The rank of the recipient of the tensor. """ if not self.stage_manager.is_last_stage(): if self.fp8_communication: cast_to_fp8_pipeline(output_tensor) self.comm.send_forward(output_tensor, next_rank, send_metadata=self.send_tensor_metadata) self.send_tensor_metadata = not self.enable_metadata_cache if self.fp8_communication: cast_from_fp8_pipeline(output_tensor, del_metadata=False) def send_backward(self, input_tensor_grad: Any, prev_rank: int = None) -> None: """Sends the gradient tensor to the previous stage in pipeline. For 1F1B. Args: input_object (Any): Object to be sent. prev_rank (int, optional): The rank of the recipient of the tensor """ if not self.stage_manager.is_first_stage(): if self.fp8_communication: cast_to_fp8_pipeline(input_tensor_grad) self.comm.send_backward(input_tensor_grad, prev_rank, send_metadata=self.send_grad_metadata) self.send_grad_metadata = not self.enable_metadata_cache if self.fp8_communication: cast_from_fp8_pipeline(input_tensor_grad, del_metadata=False) def send_forward_recv_backward(self, output_tensor: Any, send_first: Optional[bool] = None) -> Any: """Sends the input tensor to the next stage and copy the gradient tensor from the next stage in pipeline. For 1F1B. Args: output_object (Any): Object to be sent. next_rank (int, optional): The rank of the recipient of the tensor. """ if not self.stage_manager.is_last_stage(): if not self.send_tensor_metadata and self.grad_metadata_recv is not None: send_first = None if self.fp8_communication: cast_to_fp8_pipeline(output_tensor) output_tensor_grad, _ = self.comm.send_forward_recv_backward( output_tensor, send_metadata=self.send_tensor_metadata, metadata_recv=self.grad_metadata_recv, send_first=send_first, ) self.send_tensor_metadata = not self.enable_metadata_cache if self.enable_metadata_cache and self.grad_metadata_recv is None: self.grad_metadata_recv = create_send_metadata(output_tensor_grad) if self.fp8_communication: cast_from_fp8_pipeline(output_tensor, del_metadata=False) cast_from_fp8_pipeline(output_tensor_grad) return output_tensor_grad def send_backward_recv_forward(self, input_tensor_grad: Any, send_first: Optional[bool] = None) -> Any: """Sends the gradient tensor to the previous stage and copy the input tensor from the previous stage in pipeline. For 1F1B. Args: output_object (Any): Object to be sent. prev_rank (int, optional): The rank of the recipient of the tensor. """ if not self.stage_manager.is_first_stage(): if not self.send_grad_metadata and self.tensor_metadata_recv is not None: send_first = None # must not fallback if self.fp8_communication: cast_to_fp8_pipeline(input_tensor_grad) input_tensor, _ = self.comm.send_backward_recv_forward( input_tensor_grad, send_metadata=self.send_grad_metadata, metadata_recv=self.tensor_metadata_recv, send_first=send_first, ) self.send_grad_metadata = not self.enable_metadata_cache if self.enable_metadata_cache and self.tensor_metadata_recv is None: self.tensor_metadata_recv = create_send_metadata(input_tensor) if self.fp8_communication: cast_from_fp8_pipeline(input_tensor) cast_from_fp8_pipeline(input_tensor_grad, del_metadata=False) return input_tensor def forward_step( self, model: Module, input_obj: Optional[dict], criterion: Callable, accum_loss: Optional[torch.Tensor] = None, outputs: Optional[List[Any]] = None, ) -> Union[torch.Tensor, dict]: """Forward one step of the pipeline Args: model (Module): Model to be run input_obj (Optional[dict]): The output from the previous stage. If it is the first stage, the `input_obj` is None. criterion (Callable): Criterion to calculate loss. accum_loss (Optional[torch.Tensor], optional): Accumulated loss. Defaults to None. outputs (Optional[List[Any]], optional): List to store the output of the last stage (final output). Defaults to None. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ micro_batch = self.load_micro_batch() # for the first stage, input_obj is None # for the non-first stage, input_obj is the output of the previous stage and it's must be a dict output_obj = model_forward(model, micro_batch, input_obj) if self.stage_manager.is_last_stage(): loss = criterion(output_obj, micro_batch) / self.num_microbatches if accum_loss is not None: accum_loss.add_(loss.data) if outputs is not None: outputs.append(tree_map_hf(detach, output_obj)) return loss else: return output_obj def backward_step( self, optimizer: OptimizerWrapper, input_obj: Optional[dict], output_obj: Union[dict, torch.Tensor], output_obj_grad: Optional[dict], ) -> Optional[dict]: """Backward one step of the pipeline Args: optimizer (OptimizerWrapper): Optimizer to update the model input_obj (Optional[dict]): Output of the previous stage. If it is the first stage, the `input_obj` is None. output_obj (Union[dict, torch.Tensor]): Output of the current stage. If it is the last stage, the output is the loss (Tensor). output_obj_grad (dict): Gradient of the `output_obj`. If it is the last stage, the `output_obj_grad` is None. Returns: Optional[dict]: Gradient of the `input_obj`. If it is the first stage, the `input_obj_grad` is None. """ # Retain the grad on the input_obj. tree_map(retain_grad, input_obj) # Backward pass. if output_obj_grad is None: optimizer.backward(output_obj) else: keys = output_obj.get("backward_tensor_keys", output_obj_grad.keys()) tensors_to_backward = [] grads_to_backward = [] for k in keys: tensors_to_backward.append(output_obj[k]) grads_to_backward.append(output_obj_grad[k]) if len(tensors_to_backward) == 1: optimizer.backward_by_grad(tensors_to_backward[0], grads_to_backward[0]) else: optimizer.backward_by_grad(tensors_to_backward, grads_to_backward) # Collect the grad of the input_obj. input_obj_grad = None if input_obj is not None: input_obj_grad = {} for k, v in input_obj.items(): if isinstance(v, torch.Tensor) and v.grad is not None: input_obj_grad[k] = v.grad return input_obj_grad def run_forward_only( self, model: Module, data_iter: Iterable, criterion: Callable[..., Any], return_loss: bool = False, return_outputs: bool = False, ) -> Dict: """ Runs forward only schedule, with communication between pipeline stages. """ assert self.forward_only self.load_batch(data_iter) accum_loss = None if return_loss and self.stage_manager.is_last_stage(): accum_loss = torch.scalar_tensor(0, device=get_accelerator().get_current_device()) outputs = [] if return_outputs and self.stage_manager.is_last_stage() else None for _ in range(self.num_microbatches): input_obj = self.recv_forward() output_obj = self.forward_step(model, input_obj, criterion, accum_loss, outputs) self.send_forward(output_obj) if outputs is not None: if isinstance(model, ModelWrapper): model = model.unwrap() batch_size_dim = getattr(model, "batch_size_dim", 0) outputs = merge_batch(outputs, batch_size_dim) return {"loss": accum_loss, "outputs": outputs} def run_forward_backward( self, model: Module, data_iter: Iterable, criterion: Callable[..., Any], optimizer: Optional[OptimizerWrapper] = None, return_loss: bool = False, return_outputs: bool = False, ) -> Dict: """ Runs non-interleaved 1F1B schedule, with communication between pipeline stages. """ assert not self.forward_only self.load_batch(data_iter) # num_warmup_microbatches is the step when not all the processes are working num_warmup_microbatches = self.stage_manager.num_stages - self.stage_manager.stage - 1 num_warmup_microbatches = min(num_warmup_microbatches, self.num_microbatches) num_microbatches_remaining = self.num_microbatches - num_warmup_microbatches # Input, output tensors only need to be saved when doing backward passes input_objs, output_objs = [], [] accum_loss = None if return_loss and self.stage_manager.is_last_stage(): accum_loss = torch.scalar_tensor(0, device=get_current_device()) outputs = [] if return_outputs and self.stage_manager.is_last_stage() else None # Run warmup forward passes. for i in range(num_warmup_microbatches): input_obj = self.recv_forward() output_obj = self.forward_step(model, input_obj, criterion, accum_loss, outputs) self.send_forward(output_obj) input_objs.append(input_obj) output_objs.append(output_obj) # Before running 1F1B, need to receive first forward tensor. # If all microbatches are run in warmup / cooldown phase, then no need to # receive this tensor here. if num_microbatches_remaining > 0: input_obj = self.recv_forward() # Run 1F1B in steady state. for i in range(num_microbatches_remaining): last_iteration = i == (num_microbatches_remaining - 1) output_obj = self.forward_step(model, input_obj, criterion, accum_loss, outputs) output_obj_grad = self.send_forward_recv_backward(output_obj, send_first=self.stage_manager.stage % 2 == 0) # Add input_obj and output_obj to end of list. input_objs.append(input_obj) output_objs.append(output_obj) # Pop output_obj and output_obj from the start of the list for # the backward pass. input_obj = input_objs.pop(0) output_obj = output_objs.pop(0) input_obj_grad = self.backward_step(optimizer, input_obj, output_obj, output_obj_grad) if last_iteration: self.send_backward(input_obj_grad) else: input_obj = self.send_backward_recv_forward( input_obj_grad, send_first=self.stage_manager.stage % 2 == 0 ) # Run cooldown backward passes. for i in range(num_warmup_microbatches): input_obj = input_objs.pop(0) output_obj = output_objs.pop(0) output_obj_grad = self.recv_backward() input_obj_grad = self.backward_step(optimizer, input_obj, output_obj, output_obj_grad) self.send_backward(input_obj_grad) assert all(len(v) == 0 for v in input_objs) and all(len(v) == 0 for v in output_objs) if outputs is not None: if isinstance(model, ModelWrapper): model = model.unwrap() batch_size_dim = getattr(model, "batch_size_dim", 0) outputs = merge_batch(outputs, batch_size_dim) return {"loss": accum_loss, "outputs": outputs} def forward_backward_step( self, model: Module, data_iter: Iterable, criterion: Callable[..., Any], optimizer: Optional[OptimizerWrapper] = None, return_loss: bool = False, return_outputs: bool = False, ) -> dict: """ Args: model (Module): Model to be trained. data_iter (Iterable): Data iterator. criterion (Callable[[Any, Any], Tensor]): Criterion to be used. It should take two arguments: model outputs and inputs, and returns loss tensor. optimizer (OptimizerWrapper, optional): Optimizer to be used. Can be None when only forward is executed. Defaults to None. return_loss (bool, optional): Whether to return loss. Defaults to False. Whether to return loss. return_outputs (bool, optional): Whether to return model outputs. Defaults to False. Whether to return model outputs. Returns: dict: Dictionary containing loss and outputs. """ self.forward_only = not torch.is_grad_enabled() if optimizer is None: assert self.forward_only, "Optimizer should be passed when doing backward." if self.forward_only: result = self.run_forward_only(model, data_iter, criterion, return_loss, return_outputs) else: result = self.run_forward_backward(model, data_iter, criterion, optimizer, return_loss, return_outputs) return result
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/interleaved_pp.py
colossalai/pipeline/schedule/interleaved_pp.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union import torch import torch.distributed from torch.nn import Module, ModuleList from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.interface import OptimizerWrapper from colossalai.pipeline.p2p import PipelineP2PCommunication, create_send_metadata from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.quantization.fp8 import cast_from_fp8_pipeline, cast_to_fp8_pipeline from colossalai.utils import get_current_device from ._utils import detach, get_batch_size, get_micro_batch, merge_batch, model_forward, retain_grad, to_device from .base import PipelineSchedule def _wait_p2p(wait_handles) -> None: if wait_handles is not None: for req in wait_handles: req.wait() class InterleavedSchedule(PipelineSchedule): def __init__( self, stage_manager: PipelineStageManager, num_model_chunks: int, num_microbatch: Optional[int] = None, microbatch_size: Optional[int] = None, enable_metadata_cache: bool = True, overlap_p2p: bool = True, fp8_communication: bool = False, ) -> None: super().__init__(stage_manager) assert ( num_microbatch is not None or microbatch_size is not None ), "Either num_microbatch or microbatch_size should be provided" self.comm = PipelineP2PCommunication(stage_manager, overlap_p2p=overlap_p2p) self.overlap_p2p = overlap_p2p self.num_microbatch = num_microbatch self.microbatch_size = microbatch_size self.num_model_chunks = num_model_chunks self.batch: Any self.batch_size: int self.last_batch_size: Optional[int] = None self.microbatch_offset: List[int] # P2PMeta cache self.enable_metadata_cache = enable_metadata_cache self.send_tensor_metadata = True self.send_grad_metadata = True self.tensor_metadata_recv = None self.grad_metadata_recv = None self.fp8_communication = fp8_communication def load_batch(self, data_iter: Iterable, device: Optional[torch.device] = None) -> None: """Load a batch from data iterator. Args: data_iter (Iterable): Data iterator. device (Optional[torch.device], optional): Target device. Defaults to None. """ batch = next(data_iter) if device is not None: batch = tree_map(partial(to_device, device=device), batch) self.microbatch_offset = [0 for _ in range(self.num_model_chunks)] self.batch = batch self.batch_size = get_batch_size(batch) if self.microbatch_size is None: assert self.batch_size % self.num_microbatch == 0, "Batch size should divided by the number of microbatch" self.microbatch_size = self.batch_size // self.num_microbatch if self.num_microbatch is None: assert self.batch_size % self.microbatch_size == 0, "Batch size should divided by the microbatch size" self.num_microbatch = self.batch_size // self.microbatch_size if not self.forward_only: assert self.last_batch_size is None or self.last_batch_size == self.batch_size assert self.batch_size == self.microbatch_size * self.num_microbatch assert ( self.num_microbatch % self.stage_manager.num_stages == 0 ), "Number of microbatch should be an integer multiple of number of pipeline parallel devices" if self.forward_only: self.num_microbatch = (self.batch_size - 1) // self.microbatch_size + 1 # NOTE: disable metadata cache when batch size changes (not valid anymore) if self.batch_size != self.last_batch_size: self.enable_metadata_cache = False self.send_tensor_metadata = True self.send_grad_metadata = True self.tensor_metadata_recv = None self.grad_metadata_recv = None self.last_batch_size = self.batch_size def load_micro_batch(self, model_chunk_id: int) -> Any: """Load a micro batch from the current batch. Args: microbatch_id (int): the current model chunk idx. Returns: Any: Micro batch. """ assert self.microbatch_offset[model_chunk_id] <= self.batch_size, "Microbatches exhausted" micro_batch = get_micro_batch(self.batch, self.microbatch_offset[model_chunk_id], self.microbatch_size) self.microbatch_offset[model_chunk_id] += self.microbatch_size return tree_map(partial(to_device, device=get_accelerator().get_current_device()), micro_batch) def get_model_chunk_id(self, microbatch_id: int, is_forward: bool) -> int: """Helper method to get the model chunk ID given the iteration number. Args: microbatch_id (int): the current microbatch idx forward (bool): if is the forward process Returns: int: The model chunk idx of the input microbatch_id """ assert ( microbatch_id < self.num_microbatch * self.num_model_chunks ), f"microbatch_id {microbatch_id} is out of range ({self.num_microbatch * self.num_model_chunks})" microbatch_id_in_group = microbatch_id % (self.stage_manager.num_stages * self.num_model_chunks) model_chunk_id = microbatch_id_in_group // self.stage_manager.num_stages if not is_forward: # Reverse order model_chunk_id = self.num_model_chunks - model_chunk_id - 1 return model_chunk_id def recv_forward(self, model_chunk_id: int, prev_rank: int = None) -> Tuple[Any, List]: """Copy the forward output from the previous stage in pipeline as the input tensor of this stage. For interleaved 1F1B. Args: model_chunk_id (int): The current model chunk idx. prev_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input tensor or input tensor list. Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if not self.stage_manager.is_first_stage(): input_tensor, wait_handles = self.comm.recv_forward(prev_rank, metadata_recv=self.tensor_metadata_recv) if self.enable_metadata_cache and self.tensor_metadata_recv is None: self.tensor_metadata_recv = create_send_metadata(input_tensor) return input_tensor, wait_handles return None, [] def recv_backward(self, model_chunk_id: int, next_rank: int = None) -> Tuple[Any, List]: """Copy the gradient tensor from the next stage in pipeline as the input gradient of this stage. For interleaved 1F1B. Args: model_chunk_id (int): The current model chunk idx. next_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input gradient tensor or gradient tensor list. Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if not self.stage_manager.is_last_stage(): output_tensor_grad, wait_handles = self.comm.recv_backward( next_rank, metadata_recv=self.grad_metadata_recv ) if self.enable_metadata_cache and self.grad_metadata_recv is None: self.grad_metadata_recv = create_send_metadata(output_tensor_grad) return output_tensor_grad, wait_handles return None, [] def send_forward(self, model_chunk_id: int, output_tensor: Any, next_rank: int = None) -> List: """Sends the input tensor to the next stage in pipeline. For interleaved 1F1B. Args: model_chunk_id (int): The current model chunk idx. output_object (Any): Object to be sent. next_rank (int, optional): The rank of the recipient of the tensor. Returns: Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if not self.stage_manager.is_last_stage(): if self.fp8_communication: cast_to_fp8_pipeline(output_tensor) send_handles = self.comm.send_forward(output_tensor, next_rank, send_metadata=self.send_tensor_metadata) self.send_tensor_metadata = not self.enable_metadata_cache if self.fp8_communication: cast_from_fp8_pipeline(output_tensor) return send_handles return [] def send_backward(self, model_chunk_id: int, input_tensor_grad: Any, prev_rank: int = None) -> List: """Sends the gradient tensor to the previous stage in pipeline. For interleaved 1F1B. Args: model_chunk_id (int): The current model chunk idx. input_object (Any): Object to be sent. prev_rank (int, optional): The rank of the recipient of the tensor Returns: Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if not self.stage_manager.is_first_stage(): if self.fp8_communication: cast_to_fp8_pipeline(input_tensor_grad) send_handles = self.comm.send_backward( input_tensor_grad, prev_rank, send_metadata=self.send_grad_metadata ) self.send_grad_metadata = not self.enable_metadata_cache if self.fp8_communication: cast_from_fp8_pipeline(input_tensor_grad) return send_handles return [] def send_forward_recv_forward( self, model_chunk_id_send: int, model_chunk_id_recv: int, output_tensor: Any, send_first: bool = True ) -> Tuple[Any, List]: with self.stage_manager.switch_model_chunk_id(model_chunk_id_send): is_send = not self.stage_manager.is_last_stage() with self.stage_manager.switch_model_chunk_id(model_chunk_id_recv): is_recv = not self.stage_manager.is_first_stage() if self.fp8_communication: cast_to_fp8_pipeline(output_tensor) input_tensor, wait_handles = self.comm.send_forward_recv_forward( output_tensor, is_send, is_recv, send_metadata=self.send_tensor_metadata, metadata_recv=self.tensor_metadata_recv, send_first=send_first, ) # Cache metadata self.send_tensor_metadata = not self.enable_metadata_cache and is_send if is_recv and self.enable_metadata_cache and self.tensor_metadata_recv is None: self.tensor_metadata_recv = create_send_metadata(input_tensor) if self.fp8_communication: cast_from_fp8_pipeline(output_tensor) return input_tensor, wait_handles def send_backward_recv_backward( self, model_chunk_id_send: int, model_chunk_id_recv: int, input_tensor_grad: Any, send_first: bool = True ) -> Tuple[Any, List]: with self.stage_manager.switch_model_chunk_id(model_chunk_id_send): is_send = not self.stage_manager.is_first_stage() with self.stage_manager.switch_model_chunk_id(model_chunk_id_recv): is_recv = not self.stage_manager.is_last_stage() if self.fp8_communication: cast_to_fp8_pipeline(input_tensor_grad) output_tensor_grad, wait_handles = self.comm.send_backward_recv_backward( input_tensor_grad, is_send, is_recv, send_metadata=self.send_grad_metadata, metadata_recv=self.grad_metadata_recv, send_first=send_first, ) # Cache metadata self.send_grad_metadata = not self.enable_metadata_cache and is_send if is_recv and self.enable_metadata_cache and self.grad_metadata_recv is None: self.grad_metadata_recv = create_send_metadata(output_tensor_grad) if self.fp8_communication: cast_from_fp8_pipeline(input_tensor_grad) return output_tensor_grad, wait_handles def forward_step( self, model_chunk: Union[ModuleList, Module], model_chunk_id: int, input_obj: Optional[dict], criterion: Callable, accum_loss: Optional[torch.Tensor] = None, outputs: Optional[List[Any]] = None, ) -> Union[torch.Tensor, dict]: """Forward one step of the pipeline Args: model (ModuleList or Module): Model Chunk to be run input_obj (Optional[dict]): The output from the previous stage. If it is the first stage, the `input_obj` is None. criterion (Callable): Criterion to calculate loss. accum_loss (Optional[torch.Tensor], optional): Accumulated loss. Defaults to None. outputs (Optional[List[Any]], optional): List to store the output of the last stage (final output). Defaults to None. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ # Load input ids, attention mask and labels micro_batch = self.load_micro_batch(model_chunk_id=model_chunk_id) # for the first stage, input_obj is None # for other stages, input_obj is the output of the previous stage containing hidden_states etc. # Only attention_mask from micro_batch is used with self.stage_manager.switch_model_chunk_id(model_chunk_id): if isinstance(model_chunk, ModuleList): output_obj = model_forward(model_chunk[model_chunk_id], micro_batch, input_obj) else: # NOTE: in shardformer, each device still has the entire model, so we need to use relevant stage layers internal_inputs = {} if input_obj is None else input_obj internal_inputs["stage_index"] = self.stage_manager.stage_indices[model_chunk_id] output_obj = model_forward(model_chunk, micro_batch, internal_inputs) if self.stage_manager.is_last_stage(): loss = criterion(output_obj, micro_batch) / self.num_microbatch if accum_loss is not None: accum_loss.add_(loss.data) if outputs is not None: outputs.append(tree_map(detach, output_obj)) return loss else: return output_obj def backward_step( self, optimizer: OptimizerWrapper, input_obj: Optional[dict], output_obj: Union[dict, torch.Tensor], output_obj_grad: Optional[dict], ) -> Optional[dict]: """Backward one step of the pipeline Args: optimizer (OptimizerWrapper): Optimizer to update the model input_obj (Optional[dict]): Output of the previous stage. If it is the first stage, the `input_obj` is None. output_obj (Union[dict, torch.Tensor]): Output of the current stage. If it is the last stage, the output is the loss (Tensor). output_obj_grad (dict): Gradient of the `output_obj`. If it is the last stage, the `output_obj_grad` is None. Returns: Optional[dict]: Gradient of the `input_obj`. If it is the first stage, the `input_obj_grad` is None. """ # Retain the grad on the input_obj. tree_map(retain_grad, input_obj) # Backward pass. if output_obj_grad is None: optimizer.backward(output_obj) else: keys = output_obj.get("backward_tensor_keys", output_obj_grad.keys()) tensors_to_backward = [] grads_to_backward = [] for k in keys: tensors_to_backward.append(output_obj[k]) grads_to_backward.append(output_obj_grad[k]) if len(tensors_to_backward) == 1: optimizer.backward_by_grad(tensors_to_backward[0], grads_to_backward[0]) else: optimizer.backward_by_grad(tensors_to_backward, grads_to_backward) # Collect the grad of the input_obj. input_obj_grad = None if input_obj is not None: input_obj_grad = {} for k, v in input_obj.items(): if isinstance(v, torch.Tensor) and v.grad is not None: input_obj_grad[k] = v.grad return input_obj_grad def run_forward_only( self, model_chunk: Union[ModuleList, Module], data_iter: Iterable, criterion: Callable[..., Any], return_loss: bool = False, return_outputs: bool = False, ) -> Dict: assert self.forward_only self.load_batch(data_iter) outputs = [] if return_outputs and self.stage_manager.is_last_stage(ignore_chunk=True) else None accum_loss = None if return_loss and self.stage_manager.is_last_stage(ignore_chunk=True): accum_loss = torch.scalar_tensor(0, device=get_current_device()) fwd_wait_handles = [] model_chunk_id = self.get_model_chunk_id(0, is_forward=True) input_obj, fwd_wait_handles = self.recv_forward(model_chunk_id) for i in range(self.num_microbatch * self.num_model_chunks): last_batch = i == self.num_microbatch * self.num_model_chunks - 1 model_chunk_id = self.get_model_chunk_id(i, is_forward=True) # Wait until current input is received _wait_p2p(fwd_wait_handles) if self.fp8_communication and input_obj is not None: cast_from_fp8_pipeline(input_obj) output_obj = self.forward_step(model_chunk, model_chunk_id, input_obj, criterion, accum_loss, outputs) if not last_batch: input_obj, fwd_wait_handles = self.send_forward_recv_forward( model_chunk_id_send=model_chunk_id, model_chunk_id_recv=self.get_model_chunk_id(i + 1, is_forward=True), output_tensor=output_obj, send_first=self.stage_manager.stage % 2 == 0, ) else: fwd_wait_handles = self.send_forward(model_chunk_id, output_obj) if outputs is not None: outputs = merge_batch(outputs) return {"loss": accum_loss, "outputs": outputs} def run_forward_backward( self, model_chunk: Union[ModuleList, Module], data_iter: Iterable, criterion: Callable[..., Any], optimizer: Optional[OptimizerWrapper] = None, return_loss: bool = False, return_outputs: bool = False, ) -> Dict: """ Runs interleaved schedule, with communication between pipeline stages. """ assert not self.forward_only self.load_batch(data_iter) num_microbatch = self.num_microbatch * self.num_model_chunks # Forward + until 1st backward num_warmup_microbatch = (self.stage_manager.num_stages - self.stage_manager.stage - 1) * 2 # Steps needed to reach the last chunk num_warmup_microbatch += (self.num_model_chunks - 1) * self.stage_manager.num_stages num_warmup_microbatch = min(num_warmup_microbatch, num_microbatch) num_microbatch_remaining = num_microbatch - num_warmup_microbatch # Input, output tensors only need to be saved when doing backward passes input_objs = [[] for _ in range(self.num_model_chunks)] output_objs = [[] for _ in range(self.num_model_chunks)] outputs = [] if return_outputs and self.stage_manager.is_last_stage(ignore_chunk=True) else None accum_loss = None if return_loss and self.stage_manager.is_last_stage(ignore_chunk=True): accum_loss = torch.scalar_tensor(0, device=get_current_device()) bwd_wait_handles = [] # Get the 1st input batch model_chunk_id = self.get_model_chunk_id(0, is_forward=True) input_obj, fwd_wait_handles = self.recv_forward(model_chunk_id) # Run warmup forward passes. for i in range(num_warmup_microbatch): last_batch = i == num_warmup_microbatch - 1 model_chunk_id = self.get_model_chunk_id(i, is_forward=True) # Wait for input _wait_p2p(fwd_wait_handles) if self.fp8_communication and input_obj is not None: cast_from_fp8_pipeline(input_obj) output_obj = self.forward_step(model_chunk, model_chunk_id, input_obj, criterion, accum_loss, outputs) input_objs[model_chunk_id].append(input_obj) output_objs[model_chunk_id].append(output_obj) if last_batch and num_microbatch_remaining == 0: fwd_wait_handles = self.send_forward(model_chunk_id, output_obj) else: input_obj, fwd_wait_handles = self.send_forward_recv_forward( model_chunk_id_send=model_chunk_id, model_chunk_id_recv=self.get_model_chunk_id(i + 1, is_forward=True), output_tensor=output_obj, send_first=self.stage_manager.stage % 2 == 0, ) if num_microbatch_remaining > 0: model_chunk_id = self.get_model_chunk_id(0, is_forward=False) output_obj_grad, bwd_wait_handles = self.recv_backward(model_chunk_id) # Run 1F1B in steady state. for i in range(num_microbatch_remaining): fwd_batch_id = i + num_warmup_microbatch last_batch = i == num_microbatch_remaining - 1 model_chunk_id = self.get_model_chunk_id(fwd_batch_id, is_forward=True) # Wait for input. _wait_p2p(fwd_wait_handles) if self.fp8_communication and input_obj is not None: cast_from_fp8_pipeline(input_obj) output_obj = self.forward_step(model_chunk, model_chunk_id, input_obj, criterion, accum_loss, outputs) # Add input_obj and output_obj to end of list. input_objs[model_chunk_id].append(input_obj) output_objs[model_chunk_id].append(output_obj) model_chunk_id = self.get_model_chunk_id(i, is_forward=False) # Pop output_obj and output_obj from the start of the list for the backward pass. _input_obj = input_objs[model_chunk_id].pop(0) _output_obj = output_objs[model_chunk_id].pop(0) # Helper functions def send_forward_recv_forward(): if last_batch: model_chunk_id = self.get_model_chunk_id(fwd_batch_id, is_forward=True) wait_handles = self.send_forward(model_chunk_id, output_obj) return None, wait_handles else: input_obj, wait_handles = self.send_forward_recv_forward( model_chunk_id_send=self.get_model_chunk_id(fwd_batch_id, is_forward=True), model_chunk_id_recv=self.get_model_chunk_id(fwd_batch_id + 1, is_forward=True), output_tensor=output_obj, send_first=self.stage_manager.stage % 2 == 0 and i > 0, # Receive from warmup stage first in the first batch ) return input_obj, wait_handles def send_backward_recv_backward(): no_cooldown = num_microbatch == num_microbatch_remaining if last_batch and no_cooldown: model_chunk_id = self.get_model_chunk_id(i, is_forward=False) wait_handles = self.send_backward(model_chunk_id, input_obj_grad) return None, wait_handles else: output_obj_grad, wait_handles = self.send_backward_recv_backward( model_chunk_id_send=self.get_model_chunk_id(i, is_forward=False), model_chunk_id_recv=self.get_model_chunk_id(i + 1, is_forward=False), input_tensor_grad=input_obj_grad, send_first=self.stage_manager.stage % 2 == 0, ) return output_obj_grad, wait_handles input_obj, fwd_wait_handles = send_forward_recv_forward() # Wait for upstream grad _wait_p2p(bwd_wait_handles) if self.fp8_communication and output_obj_grad is not None: cast_from_fp8_pipeline(output_obj_grad) input_obj_grad = self.backward_step(optimizer, _input_obj, _output_obj, output_obj_grad) # NOTE: It's documented by NCCL that running two concurrent communicators (batch_isend_irecv) # risks deadlock (https://docs.nvidia.com/deeplearning/nccl/archives/nccl_2134/user-guide/docs/usage/communicators.html) # however in practice this works fine, and Megatron does this too # (https://github.com/microsoft/Megatron-DeepSpeed/blob/bcedecd1ff788d4d363f3365fd396053a08d65be/megatron/core/pipeline_parallel/schedules.py#L774) # if deadlock, call _wait_p2p(fwd_wait_handles) here output_obj_grad, bwd_wait_handles = send_backward_recv_backward() if num_microbatch_remaining == 0: model_chunk_id = self.get_model_chunk_id(0, is_forward=False) output_obj_grad, bwd_wait_handles = self.recv_backward(model_chunk_id) # Run cooldown backward passes. for i in range(num_microbatch_remaining, num_microbatch): last_batch = i == num_microbatch - 1 model_chunk_id = self.get_model_chunk_id(i, is_forward=False) _input_obj = input_objs[model_chunk_id].pop(0) _output_obj = output_objs[model_chunk_id].pop(0) # Wait for upstream grad _wait_p2p(bwd_wait_handles) if self.fp8_communication and output_obj_grad is not None: cast_from_fp8_pipeline(output_obj_grad) # backward local grads input_obj_grad = self.backward_step(optimizer, _input_obj, _output_obj, output_obj_grad) if not last_batch: output_obj_grad, bwd_wait_handles = self.send_backward_recv_backward( model_chunk_id_send=self.get_model_chunk_id(i, is_forward=False), model_chunk_id_recv=self.get_model_chunk_id(i + 1, is_forward=False), input_tensor_grad=input_obj_grad, send_first=self.stage_manager.stage % 2 == 0 and i > num_microbatch_remaining, ) assert (not self.overlap_p2p) or len(bwd_wait_handles) > 0 else: model_chunk_id = self.get_model_chunk_id(i, is_forward=False) _ = self.send_backward(model_chunk_id, input_obj_grad) assert all(len(v) == 0 for v in input_objs) and all(len(v) == 0 for v in output_objs) if outputs is not None: outputs = merge_batch(outputs) return {"loss": accum_loss, "outputs": outputs} def forward_backward_step( self, model_chunk: Union[ModuleList, Module], data_iter: Iterable, criterion: Callable[..., Any], optimizer: Optional[OptimizerWrapper] = None, return_loss: bool = False, return_outputs: bool = False, ) -> dict: """ Args: model_chunk (ModuleList or Module): Model Chunk to be trained. Original interleaved uses a module list whereas shardformer uses entire model + layer specification data_iter (Iterable): Data iterator. criterion (Callable[[Any, Any], Tensor]): Criterion to be used. It should take two arguments: model outputs and inputs, and returns loss tensor. optimizer (OptimizerWrapper, optional): Optimizer to be used. Can be None when only forward is executed. Defaults to None. return_loss (bool, optional): Whether to return loss. Defaults to False. Whether to return loss. return_outputs (bool, optional): Whether to return model outputs. Defaults to False. Whether to return model outputs. Returns: dict: A dict with keys: 'loss' and 'outputs'. """ self.forward_only = not torch.is_grad_enabled() if optimizer is None: assert self.forward_only, "Optimizer should be passed when doing backward." if self.forward_only: result = self.run_forward_only(model_chunk, data_iter, criterion, return_loss, return_outputs) else: result = self.run_forward_backward( model_chunk, data_iter, criterion, optimizer, return_loss, return_outputs ) return result
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/generate.py
colossalai/pipeline/schedule/generate.py
import time from functools import partial from typing import Any, Iterable, Optional, Union import torch import torch.cuda from torch.nn import Module from torch.utils._pytree import tree_map from colossalai.accelerator import get_accelerator from colossalai.inference.engine.microbatch_manager import MicroBatchManager, Status from colossalai.pipeline.p2p import PipelineP2PCommunication from colossalai.pipeline.stage_manager import PipelineStageManager from ._utils import get_batch_size, get_micro_batch, model_forward, to_device from .base import PipelineSchedule class ActionIntervalBuffer: """ The buffer to save the interval hidden states and new token for stage to use. """ def __int__(self): self.hidden_states = None self.new_token = None def clear(self): self.hidden_states = None self.new_token = None class GenerateSchedule(PipelineSchedule): """ GenerateSchedule is a class that handles the pipeline parallel inference. In our schedule, we place tie weight layer, embedding and lm_head in the same device to save space, so in this schedule, the out for each encoding progress is on rank0. Args: stage_manager (`PipelineStageManager`): Pipeline stage manager. mb_manager (`MicroBatchManager`): Micro batch manager. verbose (bool): Whether to verbose the information of the pipeline. """ def __init__(self, stage_manager: PipelineStageManager, mb_manager: MicroBatchManager, verbose: bool) -> None: super().__init__(stage_manager) self.comm = PipelineP2PCommunication(stage_manager) self.mb_manager = mb_manager self.microbatch_size = mb_manager.micro_batch_size self.batch: Optional[Any] = None self.batch_size: Optional[int] = None self.microbatch_offset: Optional[int] = None self.num_microbatches: Optional[int] = None self.action_interval_buffer = ActionIntervalBuffer() self.verbose = verbose self.timestamps = None self.comm_dtype = None def load_batch(self, data_iter: Iterable, device: Optional[torch.device] = None) -> None: """Load a batch from data iterator. Args: data_iter (Iterable): Data iterator. device (Optional[torch.device], optional): Target device. Defaults to None. """ batch = next(data_iter) if device is not None: batch = tree_map(partial(to_device, device=device), batch) self.batch = batch self.batch_size = get_batch_size(batch) if self.stage_manager.num_stages == 1: self.microbatch_size = self.batch_size self.microbatch_offset = 0 assert ( self.batch_size % self.microbatch_size == 0 ), f"Batch size should divided by the number of microbatches, {self.batch_size}, {self.num_microbatches}" self.num_microbatches = self.batch_size // self.microbatch_size self.round = self.num_microbatches // self.stage_manager.num_stages def load_micro_batch(self) -> Any: """Load a micro batch from the current batch. Returns: Any: Micro batch. """ micro_batch = get_micro_batch(self.batch, self.microbatch_offset, self.microbatch_size) self.microbatch_offset += self.microbatch_size return tree_map(partial(to_device, device=get_accelerator().get_current_device()), micro_batch) def _prepare_inputs_for_interval_stage(self): """ Prepare inputs for interval stage, for all the interval stage, the inputs is just the past_key_values Returns: dict: inputs for interval stage, `{'past_key_values': torch.Tensor}` or `None` """ model_inputs = {"infer_state": self.mb_manager.cur_description.infer_state} return model_inputs def _prepare_inputs_for_new_token(self, new_token: torch.Tensor): """ Prepare inputs for new token, the inputs is a dict with `input_ids`, `attention_mask` and `past_key_values` `input_ids` is the new token, `attention_mask` is the previous mask add `1` in the end, `past_key_values` is the past_key_values save in the micro batch manager Returns: dict: inputs for new token, `{'input_ids': torch.Tensor, 'attention_mask': torch.Tensor, 'past_key_values': torch.Tensor}` """ new_mask = self.mb_manager.cur_description.attn_mask return dict(input_ids=new_token, attention_mask=new_mask) def _get_token_id(self, hidden_state: torch.Tensor) -> torch.Tensor: last_hidden_state = hidden_state[:, -1] input_ids = torch.argmax(last_hidden_state, dim=-1).unsqueeze(1) return input_ids def _recv_pre_stage(self) -> Any: """ Receive the output from previous stage Returns: Any: The output from previous stage """ if self.stage_manager.num_stages == 2: return self.comm.p2p_recv() return self.comm.recv_forward() def _init_infer_state_action(self) -> None: """ This action is only for no first stage, to load batch and init infer_state. 1.Load micro_batch 2.Use the current micro_batch to init the current infer_state """ inputs_dict = self.load_micro_batch() self.mb_manager.add_description(inputs_dict) def _load_stage_action(self, model: Module) -> None: """ This action is only for first stage, load, init and do forward. 1.load micro_batch 2.do the forward 3.step to update """ inputs_dict = self.load_micro_batch() self.mb_manager.add_description(inputs_dict) if self.verbose and self.stage_manager.is_first_stage(): torch.cuda.synchronize() self.timestamps[self.mb_manager.idx].append(time.time()) interval_inputs = {"infer_state": self.mb_manager.cur_infer_state} output_dict = model_forward(model, inputs_dict, interval_inputs) self.action_interval_buffer.hidden_states = output_dict["hidden_states"] def _gen_token_action(self, model: Module): """ This action is only for first stage 1.do the forward with hidden_states to generate new tokens 2.step to update """ hidden_states = self.action_interval_buffer.hidden_states assert hidden_states is not None, "When first stage in GENERATE phase, the hidden states should not be None" interval_inputs = {"hidden_states": hidden_states, "infer_state": self.mb_manager.cur_infer_state} logits = model_forward(model, None, interval_inputs) if self.verbose and self.stage_manager.is_first_stage(): torch.cuda.synchronize() self.timestamps[self.mb_manager.idx].append(time.time()) assert ( "logits" in logits ), f"When first stage in GENERATE phase, the output should have attribute `logits`, but has {logits.keys()}" new_token = self._get_token_id(logits["logits"]) self.mb_manager.step(new_token) self.action_interval_buffer.new_token = new_token self.action_interval_buffer.hidden_states = None def _head_encoding_action(self, model: Module): """ In this action, 1.prepare inputs for encoding for first stage. 2.do the forward to get hidden states 3.step to update """ new_token = self.action_interval_buffer.new_token assert new_token is not None, "When first stage in GENERATE phase, the new token should not be None" inputs_dict = self._prepare_inputs_for_new_token(new_token) interval_inputs = {"infer_state": self.mb_manager.cur_infer_state} output_dict = model_forward(model, inputs_dict, interval_inputs) self.action_interval_buffer.hidden_states = output_dict["hidden_states"] def _body_encoding_action(self, model: Module): hidden_states = self.action_interval_buffer.hidden_states assert hidden_states is not None, "When not first stage, the hidden states should not be None" interval_inputs = {"hidden_states": hidden_states, "infer_state": self.mb_manager.cur_infer_state} output_dict = model_forward(model, None, interval_inputs) self.action_interval_buffer.hidden_states = output_dict["hidden_states"] def _comm_action(self, recv_pre: bool) -> torch.Tensor: """ In this action, 1.receive the hidden_states from previous stage 2.send the hidden_states to next stage """ hidden_states = self.action_interval_buffer.hidden_states ret = self.comm.p2p_communicate(hidden_states, recv_pre, comm_dtype=self.comm_dtype) self.action_interval_buffer.hidden_states = ret def _gen_action(self, model: Module): """ In p2p step method, we use `P2POp` asynchronous communication method, so the communication need to be done at the begin of each microbatch, it's a more clear way to use an action list to do so. In this function, it will generate a sequence action for current state, and do the action one by one. Args: model (Module): Model to be run. Returns: List[Callable]: A list of action, each action is a callable function, and it will be called in order. """ actions = [] if self.stage_manager.is_first_stage(): if self.mb_manager.cur_state is Status.PREFILL: actions.append(partial(self._comm_action, False)) actions.append(partial(self._load_stage_action, model)) elif self.stage_manager.is_first_stage() and self.mb_manager.cur_state is Status.GENERATE: actions.append(partial(self._comm_action, True)) actions.append(partial(self._gen_token_action, model)) actions.append(partial(self._head_encoding_action, model)) elif self.stage_manager.is_first_stage() and self.mb_manager.cur_state is Status.COOLDOWN: actions.append(partial(self._comm_action, True)) actions.append(partial(self._gen_token_action, model)) # other stage else: if self.mb_manager.cur_state is Status.PREFILL: actions.append(partial(self._init_infer_state_action)) actions.append(partial(self._comm_action, True)) actions.append(partial(self._body_encoding_action, model)) return actions def _gen_one_stage_action(self, model: Module): """ In this function, it will generate a sequence action for current state, and do the action one by one. Args: model (Module): Model to be run. Returns: List[Callable]: A list of action, each action is a callable function, and it will be called in order. """ actions = [] if self.mb_manager.cur_state is Status.PREFILL: actions.append(partial(self._load_stage_action, model)) elif self.mb_manager.cur_state is Status.GENERATE: actions.append(partial(self._gen_token_action, model)) actions.append(partial(self._head_encoding_action, model)) elif self.mb_manager.cur_state is Status.COOLDOWN: actions.append(partial(self._gen_token_action, model)) return actions def generate_step(self, model: Module, data_iter: Iterable) -> Union[torch.Tensor, dict]: if self.stage_manager.num_stages == 1: return self.generate_step_one_stage(model, data_iter) elif self.stage_manager.num_stages == 2: return self.generate_step_p2p(model, data_iter) else: return self.generate_step_broadcast(model, data_iter) @torch.no_grad() def generate_step_one_stage(self, model: Module, data_iter: Iterable) -> Union[torch.Tensor, dict]: """ Forward one step of the pipeline, when pipeline size is 1. Args: model (Module): Model to be run. data_iter (Iterable): Data iterator. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ output_sequence = [] self.load_batch(data_iter) model.eval() self.comm_dtype = model.dtype whole_timestamp = [] # run by round for _ in range(self.round): self.timestamps = [[] for _ in range(self.stage_manager.num_stages)] if self.verbose else None self.action_interval_buffer.clear() while self.mb_manager.is_micro_batch_done() is False: actions = self._gen_one_stage_action(model) for action in actions: action() self.mb_manager.next() # All microbatch in current round is DONE output_sequence.extend(self.mb_manager.export_new_tokens()) self.mb_manager.clear() if self.verbose: whole_timestamp.extend(self.timestamps) return output_sequence, whole_timestamp @torch.no_grad() def generate_step_p2p(self, model: Module, data_iter: Iterable) -> Union[torch.Tensor, dict]: """ Forward one step of the pipeline, when pipeline size is 2, the schedule is a circle, broadcast communication will be blocked, so we use `P2POp` asynchronous communication method. Args: model (Module): Model to be run. data_iter (Iterable): Data iterator. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ output_sequence = [] self.load_batch(data_iter) model.eval() self.comm_dtype = model.dtype whole_timestamp = [] # run by round for _ in range(self.round): self.timestamps = ( [[] for _ in range(self.stage_manager.num_stages)] if self.verbose and self.stage_manager.is_first_stage() else None ) self.action_interval_buffer.clear() while self.mb_manager.is_micro_batch_done() is False: actions = self._gen_action(model) for action in actions: action() self.mb_manager.next() # All microbatch in current round is DONE if self.stage_manager.is_first_stage(): output_sequence.extend(self.mb_manager.export_new_tokens()) else: self._comm_action(False) self.mb_manager.clear() if self.verbose and self.stage_manager.is_first_stage(): whole_timestamp.extend(self.timestamps) return output_sequence, whole_timestamp @torch.no_grad() def generate_step_broadcast(self, model: Module, data_iter: Iterable) -> Union[torch.Tensor, dict]: """ Forward one step of the pipeline Args: model (Module): Model to be run. data_iter (Iterable): Data iterator. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ output_sequence = [] self.load_batch(data_iter) model.eval() whole_timestamp = [] # run by round for _ in range(self.round): self.timestamps = ( [[] for _ in range(self.stage_manager.num_stages)] if self.verbose and self.stage_manager.is_first_stage() else None ) while self.mb_manager.is_micro_batch_done() is False: inputs_dict = None new_token = None output_dict = None # First stage and in PREFILL phase, just load the inputs if self.stage_manager.is_first_stage() and self.mb_manager.cur_state is Status.PREFILL: inputs_dict = self.load_micro_batch() if self.verbose and self.stage_manager.is_first_stage(): torch.cuda.synchronize() self.timestamps[self.mb_manager.idx].append(time.time()) self.mb_manager.add_description(inputs_dict) interval_inputs = {"infer_state": self.mb_manager.cur_infer_state} output_dict = model_forward(model, inputs_dict, interval_inputs) # In GENERATE phase else: # Get hidden_states from previous stage hidden_states = self.comm.recv_forward() if self.stage_manager.is_first_stage(): # First just generate a new token assert ( hidden_states is not None ), "When first stage in GENERATE phase, the hidden states should not be None" interval_inputs = { "hidden_states": hidden_states["hidden_states"], "infer_state": self.mb_manager.cur_infer_state, } logits = model_forward(model, None, interval_inputs) if self.verbose and self.stage_manager.is_first_stage(): torch.cuda.synchronize() self.timestamps[self.mb_manager.idx].append(time.time()) assert ( "logits" in logits ), f"When first stage in GENERATE phase, the output should have attribute `logits`, but has {logits.keys()}" new_token = self._get_token_id(logits["logits"]) self.mb_manager.step(new_token) # If the current micro batch is not DONE, go through blocks if self.mb_manager.cur_state in (Status.GENERATE, Status.COOLDOWN): inputs_dict = self._prepare_inputs_for_new_token(new_token) interval_inputs = {"infer_state": self.mb_manager.cur_infer_state} output_dict = model_forward(model, inputs_dict, interval_inputs) else: assert hidden_states is not None, "When not first stage, the hidden states should not be None" # inputs_dict = self._prepare_inputs_for_interval_stage() inputs_dict = None if self.mb_manager.cur_state is Status.PREFILL: inputs_dict = self.load_micro_batch() self.mb_manager.add_description(inputs_dict) interval_inputs = { "hidden_states": hidden_states["hidden_states"], "infer_state": self.mb_manager.cur_infer_state, } output_dict = model_forward(model, inputs_dict, interval_inputs) # Current microbatch is not DONE, send hidden_state to next stage if not self.stage_manager.is_first_stage() or self.mb_manager.cur_state in ( Status.GENERATE, Status.COOLDOWN, ): self.comm.send_forward({"hidden_states": output_dict["hidden_states"]}) self.mb_manager.next() # All microbatch in current round is DONE if self.stage_manager.is_first_stage(): output_sequence.extend(self.mb_manager.export_new_tokens()) self.mb_manager.clear() if self.verbose and self.stage_manager.is_first_stage(): whole_timestamp.extend(self.timestamps) return output_sequence, whole_timestamp
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/zero_bubble_pp.py
colossalai/pipeline/schedule/zero_bubble_pp.py
from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Union import torch import torch.cuda import torch.distributed from torch.nn import Module, ModuleList from torch.utils._pytree import tree_flatten, tree_map from colossalai.accelerator import get_accelerator from colossalai.interface import OptimizerWrapper from colossalai.pipeline.p2p import PipelineP2PCommunication, create_send_metadata from colossalai.pipeline.schedule.v_schedule import ScheduledNode from colossalai.pipeline.stage_manager import PipelineStageManager from colossalai.pipeline.weight_grad_store import WeightGradStore from ._utils import ( clone, detach, get_batch_size, get_micro_batch, merge_batch, model_forward, release_tensor_data, require_grad, retain_grad, to_device, ) from .base import PipelineSchedule AUTO_SCHEDULE_COMMUNICATION_TYPES = {"RECV_FORWARD", "RECV_BACKWARD", "SEND_FORWARD", "SEND_BACKWARD"} def _wait_p2p(wait_handles: List[torch.cuda.Event]) -> None: if wait_handles is not None: for req in wait_handles: req.wait() class ZeroBubbleVPipeScheduler(PipelineSchedule): r""" ZeroBubbleVPipeScheduler Args: stage_manager (PipelineStageManager): If using pipeline parallelism, it's necessary to specify a pipeline stage manager for inter-process communication in pipeline parallelism. Defaults to None, which means not using pipeline parallelism. schedule (List[ScheduledNode]): Schedule for ZeroBubbleVPipe. num_model_chunks (int) : The number of model chunk in a device. num_microbatch (Optional[int]): The number of microbatch. microbatch_size (Optional[int]): The size per microbatch. enable_metadata_cache (bool): whether to enable metadata cache to acclerate communication. overlap_p2p (bool): whether to use overlap_p2p. """ def __init__( self, stage_manager: PipelineStageManager, schedule: List[ScheduledNode], num_model_chunks: int, num_microbatch: Optional[int] = None, microbatch_size: Optional[int] = None, enable_metadata_cache: bool = True, overlap_p2p: bool = True, ): super().__init__(stage_manager) # batch info self.num_microbatch = num_microbatch self.microbatch_size = microbatch_size self.num_model_chunks = num_model_chunks self.batch: Any self.batch_size: int self.last_batch_size: Optional[int] = None self.microbatch_offset: List[int] self.schedules = schedule # TODO: optim post valid self.do_post_validation = False # P2PMeta cache self.enable_metadata_cache = enable_metadata_cache # check send_tensor_metadata, send_grad_metadata # pp4 as sample, we should follow this meta strategy # send_tensor_meta(fwd) send_grad_meta(bwd) # chunk0 | chunk1 chunk0 | chunk 1 # stage 0 T | F F | T # stage 1 T | T T | T # stage 2 T | T T | T # stage 3 F | T F | T if stage_manager.is_first_stage(ignore_chunk=True): self.send_tensor_metadata = [True, False] self.send_grad_metadata = [False, True] elif stage_manager.is_last_stage(ignore_chunk=True): self.send_tensor_metadata = [False, True] self.send_grad_metadata = [True, False] else: self.send_tensor_metadata = [True, True] self.send_grad_metadata = [True, True] # meta cache buffer self.tensor_metadata_recv = [None, None] # [chunk 0 meta, chunk 1 meta] self.grad_metadata_recv = [None, None] # P2P communication self.comm = PipelineP2PCommunication(stage_manager, overlap_p2p=overlap_p2p) # init communication map self.communication_map = { "SEND_FORWARD": self.send_forward, "RECV_FORWARD": self.recv_forward, "SEND_BACKWARD": self.send_backward, "RECV_BACKWARD": self.recv_backward, } # init buffer self._free_buffers() def _free_buffers(self): # free local buffer # two dim array, first dim is the model chunk, second dim is the microbatch queue # x & y buffer for schedule b self.input_tensors = [[], []] self.output_tensors = [[], []] # y & dy buffer for schedule w self.output_tensors_dw = [[], []] self.output_tensors_grad_dw = [[], []] # buffer for communication self.send_forward_buffer = [[], []] # [chunk0:[torch.Tensor], chunk1:[torch.Tensor]] self.recv_forward_buffer = [ [], [], ] # [chunk0:[(torch.Tensor, wait_handle)], chunk1:[(torch.Tensor, wait_handle)]] self.send_backward_buffer = [[], []] # [chunk0:[torch.Tensor], chunk1:[torch.Tensor]] self.recv_backward_buffer = [ [], [], ] # [chunk0:[(torch.Tensor, wait_handle)], chunk1:[(torch.Tensor, wait_handle)]] # y buffer for local send fwd self.local_send_forward_buffer = [] # dy buffer for local send bwd self.local_send_backward_buffer = [] # wait pp buffer self.wait_handles = [] def assert_buffer_empty(self): # assert buffer is empty at end assert len(self.input_tensors[0]) == 0 assert len(self.input_tensors[1]) == 0 assert len(self.output_tensors[0]) == 0 assert len(self.output_tensors[1]) == 0 assert len(self.output_tensors_dw[0]) == 0 assert len(self.output_tensors_dw[1]) == 0 assert len(self.output_tensors_grad_dw[0]) == 0 assert len(self.output_tensors_grad_dw[1]) == 0 assert len(self.send_forward_buffer[0]) == 0 assert len(self.send_forward_buffer[1]) == 0 assert len(self.recv_forward_buffer[0]) == 0 assert len(self.recv_forward_buffer[1]) == 0 assert len(self.send_backward_buffer[0]) == 0 assert len(self.send_backward_buffer[1]) == 0 assert len(self.recv_backward_buffer[0]) == 0 assert len(self.recv_backward_buffer[1]) == 0 assert len(self.local_send_forward_buffer) == 0 assert len(self.local_send_backward_buffer) == 0 def load_batch(self, data_iter: Iterable, device: Optional[torch.device] = None) -> None: """Load a batch from data iterator. Args: data_iter (Iterable): Data iterator. device (Optional[torch.device], optional): Target device. Defaults to None. """ batch = next(data_iter) if device is not None: batch = tree_map(partial(to_device, device=device), batch) self.microbatch_offset = [0 for _ in range(self.num_model_chunks)] self.batch = batch self.batch_size = get_batch_size(batch) if self.microbatch_size is None: assert self.batch_size % self.num_microbatch == 0, "Batch size should divided by the number of microbatch" self.microbatch_size = self.batch_size // self.num_microbatch if self.num_microbatch is None: assert self.batch_size % self.microbatch_size == 0, "Batch size should divided by the microbatch size" self.num_microbatch = self.batch_size // self.microbatch_size if not self.forward_only: assert self.last_batch_size is None or self.last_batch_size == self.batch_size assert self.batch_size == self.microbatch_size * self.num_microbatch assert ( self.num_microbatch % self.stage_manager.num_stages == 0 ), "Number of microbatch should be an integer multiple of number of pipeline parallel devices" if self.forward_only: self.num_microbatch = (self.batch_size - 1) // self.microbatch_size + 1 self.last_batch_size = self.batch_size def load_micro_batch(self, model_chunk_id: int) -> Any: """Load a micro batch from the current batch. Args: microbatch_id (int): the current model chunk idx. Returns: Any: Micro batch. """ assert self.microbatch_offset[model_chunk_id] <= self.batch_size, "Microbatches exhausted" micro_batch = get_micro_batch(self.batch, self.microbatch_offset[model_chunk_id], self.microbatch_size) self.microbatch_offset[model_chunk_id] += self.microbatch_size return tree_map(partial(to_device, device=get_accelerator().get_current_device()), micro_batch) def get_model_chunk_id(self, microbatch_id: int, is_forward: bool) -> int: """Helper method to get the model chunk ID given the iteration number. Args: microbatch_id (int): the current microbatch idx forward (bool): if is the forward process Returns: int: The model chunk idx of the input microbatch_id """ assert ( microbatch_id < self.num_microbatch * self.num_model_chunks ), f"microbatch_id {microbatch_id} is out of range ({self.num_microbatch * self.num_model_chunks})" microbatch_id_in_group = microbatch_id % (self.stage_manager.num_stages * self.num_model_chunks) model_chunk_id = microbatch_id_in_group // self.stage_manager.num_stages if not is_forward: # Reverse order model_chunk_id = self.num_model_chunks - model_chunk_id - 1 return model_chunk_id def recv_forward(self, model_chunk_id: int, prev_rank: int = None) -> List: """Copy the forward output from the previous stage in pipeline as the input tensor of this stage. For ZBV. Args: model_chunk_id (int): The current model chunk idx. prev_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input tensor or input tensor list. Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if model_chunk_id == 0: ################ # chunk = 0 & is_first_stage # do nothing; cause u are chunk 0 in first rank, u have no prev rank; ################# if self.stage_manager.is_first_stage(ignore_chunk=True): return [] ################ # chunk = 0 & not is_first_stage # Recv y from PREV_rank as input ################# else: prev_rank = self.stage_manager.get_prev_rank() input_tensor, wait_handles = self.comm.recv_forward( prev_rank=prev_rank, metadata_recv=self.tensor_metadata_recv[model_chunk_id] ) if self.enable_metadata_cache and self.tensor_metadata_recv[model_chunk_id] is None: self.tensor_metadata_recv[model_chunk_id] = create_send_metadata(input_tensor) self.recv_forward_buffer[model_chunk_id].append((input_tensor, wait_handles)) return wait_handles else: ################ # chunk = 1 & is_last_stage # do nothing; cause u get y from local_send_forward_buffer in schedule f ################ if self.stage_manager.is_last_stage(ignore_chunk=True): # return None, [] return [] ################ # chunk = 1 & not is_last_stage # recv y from NEXT_rank as input ################ else: next_rank = self.stage_manager.get_next_rank() input_tensor, wait_handles = self.comm.recv_forward( next_rank, metadata_recv=self.tensor_metadata_recv[model_chunk_id] ) if self.enable_metadata_cache and self.tensor_metadata_recv[model_chunk_id] is None: self.tensor_metadata_recv[model_chunk_id] = create_send_metadata(input_tensor) self.recv_forward_buffer[model_chunk_id].append((input_tensor, wait_handles)) return wait_handles def recv_backward(self, model_chunk_id: int, next_rank: int = None) -> List: """Copy the gradient tensor from the next stage in pipeline as the input gradient of this stage. For ZBV. Args: model_chunk_id (int): The current model chunk idx. next_rank (int, optional): The rank of the source of the tensor. Returns: Any: The input gradient tensor or gradient tensor list. Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if model_chunk_id == 0: # bwd chunk0 is right V; ################ # chunk = 0 & is_last_stage # do nothing; Already get dy from local_send_backward_buffer in schedule b ################ if self.stage_manager.is_last_stage(ignore_chunk=True): return [] ################ # chunk = 0 & not is_last_stage # Recv bwd from next stage; ################ else: next_rank = self.stage_manager.get_next_rank() output_tensor_grad, wait_handles = self.comm.recv_backward( next_rank, metadata_recv=self.grad_metadata_recv[model_chunk_id] ) if self.enable_metadata_cache and self.grad_metadata_recv[model_chunk_id] is None: self.grad_metadata_recv[model_chunk_id] = create_send_metadata(output_tensor_grad) self.recv_backward_buffer[model_chunk_id].append((output_tensor_grad, wait_handles)) return wait_handles else: # bwd chunk1 is left V; ################ # chunk = 1 & is_first_stage # do nothing; get loss from local ################ if self.stage_manager.is_first_stage(ignore_chunk=True): return [] ################ # chunk = 1 & not first stage # recv_backward recv bwd from prev stage; ################ else: prev_rank = self.stage_manager.get_prev_rank() output_tensor_grad, wait_handles = self.comm.recv_backward( next_rank=prev_rank, metadata_recv=self.grad_metadata_recv[model_chunk_id] ) if self.enable_metadata_cache and self.grad_metadata_recv[model_chunk_id] is None: self.grad_metadata_recv[model_chunk_id] = create_send_metadata(output_tensor_grad) self.recv_backward_buffer[model_chunk_id].append((output_tensor_grad, wait_handles)) return wait_handles def send_forward(self, model_chunk_id: int, next_rank: int = None) -> List: """Sends the input tensor to the next stage in pipeline. For ZBV. Args: model_chunk_id (int): The current model chunk idx. next_rank (int, optional): The rank of the recipient of the tensor. Returns: Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if model_chunk_id == 0: ################ # chunk = 0 && is_last_stage # do nothing; hold y on local_send_forward_buffer ################ if self.stage_manager.is_last_stage(ignore_chunk=True): self.send_tensor_metadata[model_chunk_id] = not self.enable_metadata_cache return [] ################ # chunk = 0 && not is_last_stage # self.comm.send_forward send y to NEXT stage ################ else: next_rank = self.stage_manager.get_next_rank() output_tensor = self.send_forward_buffer[model_chunk_id].pop(0) send_handles = self.comm.send_forward( output_object=output_tensor, next_rank=next_rank, send_metadata=self.send_tensor_metadata[model_chunk_id], ) self.send_tensor_metadata[model_chunk_id] = not self.enable_metadata_cache return send_handles else: ################ # chunk = 1 && is_first_stage # do nothing; Already send LOSS to local_send_backward_buffer in schedule f send part ################ if self.stage_manager.is_first_stage(ignore_chunk=True): self.send_tensor_metadata[model_chunk_id] = not self.enable_metadata_cache return [] ################ # chunk = 1 && not is_first_stage # self.comm.send_forward send y to PREV stage ################ else: prev_rank = self.stage_manager.get_prev_rank() output_tensor = self.send_forward_buffer[model_chunk_id].pop(0) send_handles = self.comm.send_forward( output_tensor, prev_rank, send_metadata=self.send_tensor_metadata[model_chunk_id] ) self.send_tensor_metadata[model_chunk_id] = not self.enable_metadata_cache return send_handles def send_backward(self, model_chunk_id: int, prev_rank: int = None) -> List: """Sends the gradient tensor to the previous stage in pipeline. For ZBV. Args: model_chunk_id (int): The current model chunk idx. prev_rank (int, optional): The rank of the recipient of the tensor Returns: Any: The wait handles for the communication. """ with self.stage_manager.switch_model_chunk_id(model_chunk_id): if model_chunk_id == 0: # bwd chunk0 is right V; ################ # chunk = 0 && is_first_stage # do nothing; cause u are the first chunk in first stage; bwd end ################ if self.stage_manager.is_first_stage(ignore_chunk=True): self.send_grad_metadata[model_chunk_id] = not self.enable_metadata_cache return [] ################ # chunk = 0 && not is_first_stage # Send dx to PREV stage; ################ else: prev_rank = self.stage_manager.get_prev_rank() input_tensor_grad = self.send_backward_buffer[model_chunk_id].pop(0) send_handles = self.comm.send_backward( input_tensor_grad, prev_rank, send_metadata=self.send_grad_metadata[model_chunk_id] ) self.send_grad_metadata[model_chunk_id] = not self.enable_metadata_cache return send_handles # bwd chunk1 is left V; else: ################ # chunk = 1 && is_last_stage # do nothing; Already send input_tensor_grad to local_send_bwd_buffer in schedule b; ################ if self.stage_manager.is_last_stage(ignore_chunk=True): self.send_grad_metadata[model_chunk_id] = not self.enable_metadata_cache return [] ################ # chunk = 1 && not is_last_stage # Send dx to NEXT stage; ################ else: next_rank = self.stage_manager.get_next_rank() input_tensor_grad = self.send_backward_buffer[model_chunk_id].pop(0) send_handles = self.comm.send_backward( input_tensor_grad, next_rank, send_metadata=self.send_grad_metadata[model_chunk_id] ) self.send_grad_metadata[model_chunk_id] = not self.enable_metadata_cache return send_handles def forward_step( self, model_chunk: Union[ModuleList, Module], model_chunk_id: int, micro_batch: Optional[dict], input_obj: Optional[dict], criterion: Callable, accum_loss: Optional[torch.Tensor] = None, outputs: Optional[List[Any]] = None, ) -> Union[torch.Tensor, dict]: """Forward one step of the pipeline Args: model_chunk (ModuleList or Module): Model Chunk to be run; model_chunk_id (int): The current model chunk idx; input_obj (Optional[dict]): x; criterion (Callable): loss function; accum_loss (Optional[torch.Tensor], optional): Accumulated loss. Defaults to None. outputs (Optional[List[Any]], optional): List to store the output of the last stage (final output). Defaults to None. Returns: Union[torch.Tensor, dict]: The intermediate output (dict) of the current stage. If it is the last stage, the output is the loss (Tensor). """ # Load input ids, attention mask and labels # for the first stage, input_obj is None; So,we use micro_batch as input_obj # for other stages, input_obj is the output of the previous/next stage containing hidden_states etc. # Only attention_mask from micro_batch is used with self.stage_manager.switch_model_chunk_id(model_chunk_id): # fwd calculate internal_inputs = {} if input_obj is None else input_obj internal_inputs["stage_index"] = self.stage_manager.stage_indices[model_chunk_id] output_obj = model_forward(model_chunk, micro_batch, internal_inputs) # last layer in model if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): loss = criterion(output_obj, micro_batch) / self.num_microbatch if accum_loss is not None: accum_loss.add_(loss.detach()) if outputs is not None: outputs.append(tree_map(detach, output_obj)) return loss else: return output_obj def backward_b_step( self, model_chunk: Union[ModuleList, Module], model_chunk_id: int, optimizer: OptimizerWrapper, # micro_batch: Optional[dict], input_obj: Optional[dict], output_obj: Union[dict, torch.Tensor], output_obj_grad: Optional[dict], ) -> Optional[dict]: """Backward dx step of the pipeline; we calculate "dx = w*dy" here; Args: model_chunk (ModuleList or Module): Model Chunk to be run; model_chunk_id (int): The current model chunk idx; optimizer (OptimizerWrapper): Optimizer to update the model input_obj (Optional[Tuple(dict)]): x. (microbatch, input_obj) output_obj (Union[dict, torch.Tensor]): y. output_obj_grad (dict): dy. Returns: Optional[dict]: dx. """ # calculate bwd b step ; only dx = w*dy; # Retain the grad on the input_obj. No need retain_grad microbatch if input_obj is not None: tree_map(retain_grad, input_obj) # x, y, dy list for backward_by_grad; Type: list[tensor]; input_obj_ = [] output_obj_ = [] output_obj_grad_ = [] # For chunk 0 stage 0, use micro_batch as input_obj_; and we don't have to cal microbatch dx. # For loss backward; output_obj is loss; output_obj_grad should be None if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): assert output_obj_grad is None input_obj_, _ = tree_flatten(input_obj) output_obj_.append(output_obj) # LOSS output_obj_grad_.append(output_obj_grad) # None # For other chunk stage, use input_obj as input_obj_; else: input_obj_, _ = tree_flatten(input_obj) output_obj_, _ = tree_flatten(output_obj) # y output_obj_grad_, _ = tree_flatten(output_obj_grad) # dy # filter item which is not torch.Tensor input_obj_ = [v for v in input_obj_ if isinstance(v, torch.Tensor) or v is None] output_obj_ = [v for v in output_obj_ if isinstance(v, torch.Tensor) or v is None] output_obj_grad_ = [v for v in output_obj_grad_ if isinstance(v, torch.Tensor) or v is None] try: ctx = optimizer.no_sync() except AttributeError: ctx = model_chunk.no_sync() with ctx: optimizer.backward_by_grad( tensor=output_obj_, grad=output_obj_grad_, # inputs=input_obj_, retain_graph=False, ) # Format output_obj_grad input_obj_grad = dict() if model_chunk_id == 0 and self.stage_manager.is_first_stage(ignore_chunk=True): pass else: for k, v in input_obj.items(): if isinstance(v, torch.Tensor) and v.grad is not None: input_obj_grad[k] = v.grad return input_obj_grad def backward_w_step( self, model_chunk: Union[ModuleList, Module], model_chunk_id: int, optimizer: OptimizerWrapper, output_obj: Union[dict, torch.Tensor], output_obj_grad: Optional[dict], ): """Backward dw step of the pipeline; we calculate "dw = x*dy" here; Args: model_chunk (ModuleList or Module): Model Chunk to be run; model_chunk_id (int): The current model chunk idx; optimizer (OptimizerWrapper): Optimizer to update the model output_obj (Union[dict, torch.Tensor]): y. output_obj_grad (dict): dy. Returns: Nothing need to return; we only calculate dw then update w; """ # calculate bwd w step ; only dw = x*dy; # y, dy list for w backward_by_grad; Type: list[tensor]; output_obj_ = [] output_obj_grad_ = [] if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): # loss backward; output_obj is loss; output_obj_.append(output_obj) # LOSS output_obj_grad_.append(None) # None else: output_obj_, _ = tree_flatten(output_obj) # y output_obj_grad_, _ = tree_flatten(output_obj_grad) # dy # filter item which is not torch.Tensor output_obj_ = [v for v in output_obj_ if isinstance(v, torch.Tensor) or v is None] output_obj_grad_ = [v for v in output_obj_grad_ if isinstance(v, torch.Tensor) or v is None] optimizer.backward_by_grad( tensor=output_obj_, grad=output_obj_grad_, inputs=list(model_chunk.parameters()), retain_graph=False, ) def schedule_f( self, scheduled_node, model_chunk: torch.nn.ModuleList, model_chunk_id: int, criterion: Callable, accum_loss: Optional[torch.Tensor] = None, outputs: Optional[List[Any]] = None, ): """A complete forward schedule; Include recv fwd --> cal fwd --> send fwd; Args: scheduled_node: model_chunk (ModuleList or Module): Model Chunk to be run; model_chunk_id (int): The current model chunk idx; criterion (Callable): loss function; accum_loss (Optional[torch.Tensor], optional): Accumulated loss. Defaults to None. outputs (Optional[List[Any]], optional): List to store the output of the last stage (final output). Defaults to None. Returns: Nothing. """ micro_batch = self.load_micro_batch(model_chunk_id=model_chunk_id) # Step1: recv fwd if model_chunk_id == 0: # is first stage; get input from microbatch if self.stage_manager.is_first_stage(ignore_chunk=True): input_obj = None # (tensor, wait_handle) else: input_obj = self.recv_forward_buffer[model_chunk_id].pop(0) for h in input_obj[1]: h.wait() input_obj = input_obj[0] else: # is last stage; recv from local if self.stage_manager.is_last_stage(ignore_chunk=True): input_obj = self.local_send_forward_buffer.pop(0) # not last stage; recv from next else: input_obj = self.recv_forward_buffer[model_chunk_id].pop(0) for h in input_obj[1]: h.wait() input_obj = input_obj[0] # Here, let input_obj.requires_grad_() # if input_obj is not None: if not isinstance(input_obj, torch.Tensor): tree_map(require_grad, input_obj) # Also requires_grad_ for micro_batch in stage 0 chunk 0 fwd, # tree_map(torch.Tensor.requires_grad_, micro_batch) # Step2: fwd step output_obj = self.forward_step( model_chunk=model_chunk, model_chunk_id=model_chunk_id, micro_batch=micro_batch, input_obj=input_obj, criterion=criterion, accum_loss=accum_loss, outputs=outputs, ) # Step3: # 3-1:detach output; detach output for send fwd; if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): # We should not detach bwd LOSS pass else: # detach output detached_output_obj = tree_map(detach, output_obj) # 3-2 clone detached_output_obj detached_output_obj = tree_map(clone, detached_output_obj) # 3-3 release cloned output.data; release_tensor_data output for bwd b & w; (do not detach output) if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): # We should not release_tensor_data bwd LOSS pass else: # release_tensor_data output tree_map(release_tensor_data, output_obj) # add input and output object for backward b self.input_tensors[model_chunk_id].append(input_obj) # for bwd b&w, we only need the graph(grad_fn) of output_obj # Do not release_tensor_data loss, release_tensor_data other output_obj; if model_chunk_id == 1 and self.stage_manager.is_first_stage(ignore_chunk=True): self.output_tensors[model_chunk_id].append(output_obj) else: self.output_tensors[model_chunk_id].append(output_obj) # add output to send_fwd_buffer if model_chunk_id == 0: # chunk 0 # is last stage; send to local_send_forward_buffer if self.stage_manager.is_last_stage(ignore_chunk=True): self.local_send_forward_buffer.append(detached_output_obj) else: self.send_forward_buffer[model_chunk_id].append(detached_output_obj) else: # chunk 1 # is first stage; end of fwd; do nothing if self.stage_manager.is_first_stage(ignore_chunk=True): pass else: self.send_forward_buffer[model_chunk_id].append(detached_output_obj) def schedule_b( self, scheduled_node, model_chunk: Union[ModuleList, Module], model_chunk_id: int, optimizer: OptimizerWrapper, ): """A complete backward b schedule; Include recv bwd --> cal bwd step --> send bwd; Args: scheduled_node: model_chunk (ModuleList or Module): Model Chunk to be run; model_chunk_id (int): The current model chunk idx; Returns: Nothing. """ # Step1: recv bwd
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/__init__.py
colossalai/pipeline/schedule/__init__.py
from .base import PipelineSchedule from .interleaved_pp import InterleavedSchedule from .one_f_one_b import OneForwardOneBackwardSchedule from .zero_bubble_pp import ZeroBubbleVPipeScheduler __all__ = [ "PipelineSchedule", "OneForwardOneBackwardSchedule", "InterleavedSchedule", "ZeroBubbleVPipeScheduler", ]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/base.py
colossalai/pipeline/schedule/base.py
from typing import Any, Callable, Iterable, Optional from torch import Tensor from torch.nn import Module from colossalai.interface import OptimizerWrapper from colossalai.pipeline.stage_manager import PipelineStageManager class PipelineSchedule: def __init__(self, stage_manager: PipelineStageManager) -> None: self.stage_manager = stage_manager def forward_backward_step( self, model: Module, data_iter: Iterable, criterion: Callable[[Any, Any], Tensor], optimizer: Optional[OptimizerWrapper] = None, return_loss: bool = False, return_outputs: bool = False, ) -> dict: """Forward and backward step for pipeline training. Args: model (Module): Model to be trained. data_iter (Iterable): Data iterator. criterion (Callable[[Any, Any], Tensor]): Criterion to be used. It should take two arguments: model outputs and inputs, and returns loss tensor. optimizer (OptimizerWrapper, optional): Optimizer to be used. Can be None when only forward is executed. Defaults to None. return_loss (bool, optional): Whether to return loss. Defaults to False. Whether to return loss. return_outputs (bool, optional): Whether to return model outputs. Defaults to False. Whether to return model outputs. Returns: dict: A dict with keys: 'loss' and 'outputs'. """ raise NotImplementedError
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/v_schedule.py
colossalai/pipeline/schedule/v_schedule.py
# Refer from Zero Bubble Pipeline Parallelism. # Github: https://github.com/sail-sg/zero-bubble-pipeline-parallelism # Paper: https://arxiv.org/abs/2401.10241 # The following applies to all files unless otherwise noted: # Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of NVIDIA CORPORATION nor the names of its # contributors may be used to endorse or promote products derived # from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY # EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR # PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR # CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY # OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. from collections import deque from dataclasses import dataclass @dataclass(eq=True, frozen=True) class ScheduledNode: type: str chunk: int stage: int minibatch: int start_time: int = 0 completion_time: int = 0 rollback: bool = False class PipelineGraph(object): """PipelineGraph""" def __init__( self, n_stage, n_micro, f_cost, b_cost, w_cost, c_cost, f_mem, b_mem, w_mem, max_mem=None, ): self.n_node = 6 * n_stage * n_micro self.n_stage = n_stage self.n_micro = n_micro self.f_cost = f_cost self.b_cost = b_cost self.w_cost = w_cost self.c_cost = c_cost self.f_mem = f_mem self.b_mem = b_mem self.w_mem = w_mem self.fbw_cost = [f_cost, b_cost, w_cost] self.fbw_mem = [f_mem, b_mem, w_mem] self.max_mem = max_mem or f_mem * self.n_stage * 2 def get_id(self, cat, chunk, stage, micro): return ( cat * 2 * self.n_stage * self.n_micro + chunk * self.n_stage * self.n_micro + stage * self.n_micro + micro ) def try_v_schedule(self, fill_f=True, fill_b=True, approved_bubble=None): count = [] for i in range(self.n_stage): count.append([0] * 6) end_time = [-1] * self.n_node cur_time = [0] * self.n_stage mem = [0] * self.n_stage stage_bubble = [0] * self.n_stage pending_w = [deque() for _ in range(self.n_stage)] schedule = [[] for _ in range(self.n_stage)] stage_str = [" " * i for i in range(self.n_stage)] if approved_bubble is None: approved_bubble = [-1] * self.n_stage max_approved_bubble = max(approved_bubble) def get_max_stage_bubble(stage=-1): max_stage_bubble = 0 for bb in stage_bubble: max_stage_bubble = max(max_stage_bubble, bb) if stage >= 0: max_stage_bubble = max(max_stage_bubble, max_approved_bubble - approved_bubble[stage]) return max_stage_bubble def put_w(stage): assert len(pending_w[stage]) > 0 _, chunk_, _ = pending_w[stage].popleft() put(2, chunk_, stage) def put(cat, chunk, stage, assert_cnt=True): _tmp = _no_bubble = cur_time[stage] + self.fbw_cost[cat] _cnt = count[stage][cat * 2 + chunk] # assert _cnt < self.n_micro if _cnt >= self.n_micro: if not assert_cnt: stage_str[stage] += " " cur_time[stage] = _tmp # TODO return assert False assert mem[stage] + self.fbw_mem[cat] <= self.max_mem stage_str[stage] += "FfBbWw"[cat * 2 + chunk] + str(_cnt + 1) + " " * (3 - len(str(_cnt + 1))) if cat > 0 or chunk > 0: last_id = cat * 2 + chunk - 1 if cat < 2: assert end_time[self.get_id(last_id // 2, last_id % 2, stage, _cnt)] >= 0 else: assert end_time[self.get_id(1, chunk, stage, _cnt)] >= 0 if chunk == 1 and cat < 2: if stage < self.n_stage - 1: _fa_id = self.get_id(cat, chunk, stage + 1, _cnt) assert end_time[_fa_id] >= 0 _tmp = max(_tmp, end_time[_fa_id] + self.c_cost + self.fbw_cost[cat]) if chunk == 0 and cat < 2: if stage > 0: _fa_id = self.get_id(cat, chunk, stage - 1, _cnt) assert end_time[_fa_id] >= 0, f"{cat}, {chunk}, {stage}, {_cnt}" _tmp = max(_tmp, end_time[_fa_id] + self.c_cost + self.fbw_cost[cat]) _id = self.get_id(cat, chunk, stage, _cnt) if count[stage][0] > 0: stage_bubble[stage] += _tmp - _no_bubble end_time[_id] = _tmp cur_time[stage] = _tmp mem[stage] += self.fbw_mem[cat] # noinspection PyTypeChecker schedule[stage].append((cat, chunk, _cnt)) if cat == 1: pending_w[stage].append((2, chunk, _cnt)) count[stage][cat * 2 + chunk] += 1 for i in range(self.n_stage): put(0, 0, i) for i in range(self.n_stage - 1, -1, -1): if i == self.n_stage - 1: put(0, 1, i) continue tmp = end_time[self.get_id(0, 1, i + 1, 0)] + self.c_cost while ( mem[i] + self.fbw_mem[0] * (2 + i * 2) <= self.max_mem and cur_time[i] + self.fbw_cost[0] <= tmp and count[i][0] < self.n_micro ): for j in range(i + 1): put(0, 0, j) put(0, 1, i) iter_chunk_ = 0 end_tmp = 0 for i in range(self.n_stage): if i == 0: end_tmp = cur_time[0] + self.fbw_cost[1] continue tmp = end_tmp + self.c_cost while ( count[i][0] + count[i][1] < count[i - 1][0] + count[i - 1][1] or count[i][1] <= count[i - 1][1] < self.n_micro ): for j in range(self.n_stage - 1, i - 1, -1): if count[j][iter_chunk_] < self.n_micro: put(0, iter_chunk_, j) iter_chunk_ = 1 - iter_chunk_ for _ in range(2 * self.n_micro): # check mem before putting b for i in range(self.n_stage): while mem[i] + self.fbw_mem[1] > self.max_mem: assert len(pending_w[i]) > 0 put_w(i) b0_ranks, b1_ranks = [], [] for i in range(self.n_stage): if count[i][3] >= count[i][2]: b0_ranks.append(i) elif i == self.n_stage - 1: b1_ranks.append(i) else: fa_id = self.get_id(1, 1, i + 1, count[i][3]) if end_time[fa_id] >= 0 or count[i][2] >= self.n_micro: b1_ranks.append(i) else: b0_ranks.append(i) b_ranks = [] # put b1 for i in reversed(b1_ranks): b_ranks.append((i, 1)) # put b0 for i in b0_ranks: b_ranks.append((i, 0)) for i, _chunk_ in b_ranks: fa_id = -1 if _chunk_ == 1 and i < self.n_stage - 1: fa_id = self.get_id(1, 1, i + 1, count[i][3]) if _chunk_ == 0 and i > 0: fa_id = self.get_id(1, 0, i - 1, count[i][2]) while ( len(pending_w[i]) > 0 and fa_id >= 0 and end_time[fa_id] + self.c_cost >= cur_time[i] + self.fbw_cost[2] ): # fill the bubble put_w(i) if ( len(pending_w[i]) > 0 and end_time[fa_id] + self.c_cost - cur_time[i] > get_max_stage_bubble(i) - stage_bubble[i] ): if _chunk_ == 1: put_w(i) elif fill_b: put_w(i) put(1, _chunk_, i) # put f for i in range(self.n_stage): if count[i][1] >= self.n_micro: continue put_item = None if count[i][1] >= count[i][0]: put_item = 0 elif i == self.n_stage - 1: put_item = 1 else: if end_time[self.get_id(0, 1, i + 1, count[i][1])] >= 0: put_item = 1 elif count[i][0] < self.n_micro: if i == 0: put_item = 0 elif end_time[self.get_id(0, 0, i - 1, count[i][0])] >= 0: put_item = 0 if put_item is None: continue # check mem before putting f while mem[i] + self.fbw_mem[0] > self.max_mem: assert len(pending_w[i]) > 0 put_w(i) fa_id = -1 if put_item == 0 and i > 0: fa_id = self.get_id(0, 0, i - 1, count[i][0]) if put_item == 1 and i < self.n_stage - 1: fa_id = self.get_id(0, 1, i + 1, count[i][1]) while ( len(pending_w[i]) > 0 and fa_id >= 0 and end_time[fa_id] + self.c_cost >= cur_time[i] + self.fbw_cost[2] ): # fill the bubble put_w(i) if ( len(pending_w[i]) > 0 and end_time[fa_id] + self.c_cost - cur_time[i] > get_max_stage_bubble(i) - stage_bubble[i] ): if fill_f: put_w(i) put(0, put_item, i) for i in range(self.n_stage): while len(pending_w[i]) > 0: put_w(i) max_bubble = get_max_stage_bubble() expected_time = sum(self.fbw_cost) * self.n_micro * 2 max_bubble / expected_time if max_approved_bubble < 0 or max_bubble < max_approved_bubble: _schedule, _end_time, _max_bubble = self.try_v_schedule( fill_f=fill_f, fill_b=fill_b, approved_bubble=stage_bubble, ) if _max_bubble < max_bubble: return _schedule, _end_time, _max_bubble return schedule, end_time, max_bubble def print_details(self, end_time, print_scaling=1): for stage in range(self.n_stage): stage_str = ["."] * int(max(end_time) / print_scaling) for _cat in range(3): for _chunk in range(2): for _micro in range(self.n_micro): _id = self.get_id(_cat, _chunk, stage, _micro) if end_time[_id] < 0: continue end = int(end_time[_id] / print_scaling) start = int((end_time[_id] - self.fbw_cost[_cat]) / print_scaling) for j in range(start, end): if j == start or j == end - 1: stage_str[j] = "FfBbWw"[_cat * 2 + _chunk] elif j == start + 1: if _micro >= 10: stage_str[j] = str(_micro // 10) else: stage_str[j] = str(_micro) elif j == start + 2 and _micro >= 10: stage_str[j] = str(_micro % 10) else: stage_str[j] = "-" _str = "" for _c in stage_str: _str += _c print(_str) def get_v_schedule(self, only_run_time=False): schedule, end_time, max_bubble = None, None, None expected_time = sum(self.fbw_cost) * self.n_micro * 2 for fill_b in [True, False]: for fill_f in [True, False]: _schedule, _end_time, _max_bubble = self.try_v_schedule(fill_b=fill_b, fill_f=fill_f) if max_bubble is None or _max_bubble < max_bubble: max_bubble = _max_bubble schedule = _schedule end_time = _end_time if only_run_time: return max_bubble + expected_time max_bubble / (expected_time + max_bubble) local_order = [[] for _ in range(self.n_stage)] comm_id = {} comm_id_counter = 0 post_validation_time = 0 for i in range(self.n_stage - 1, -1, -1): pv_id = min(2 * (self.n_stage - 1 - i), self.n_micro - 1) post_validation_time = max( post_validation_time, end_time[self.get_id(0, 0, i, pv_id)] - self.fbw_cost[0] - self.c_cost ) # post_validation_time = 0 for it in ["RECV_", "SEND_", ""]: if i == 0 and it == "SEND_": continue if i == self.n_stage - 1 and it == "RECV_": continue # stage_ = i - 1 if it == "RECV_" else i stage_ = i local_order[stage_].append( ScheduledNode( type=it + "POST_VALIDATION", chunk=0, stage=stage_, minibatch=0, start_time=post_validation_time, completion_time=post_validation_time, ) ) comm_id[local_order[stage_][-1]] = comm_id_counter comm_id_counter += 1 for i in range(self.n_stage): for _cat_, _chunk_, _micro_ in schedule[i]: complete_time = end_time[self.get_id(_cat_, _chunk_, i, _micro_)] local_order[i].append( ScheduledNode( type="FBW"[_cat_], chunk=_chunk_ if _cat_ == 0 else 1 - _chunk_, stage=i, minibatch=_micro_, start_time=complete_time - self.fbw_cost[_cat_], completion_time=complete_time, ) ) if _cat_ == 2: # no communication for W continue cat_str = "FORWARD" if _cat_ == 0 else "BACKWARD" def communicate(send_recv, stage_): # noinspection PyTypeChecker local_order[stage_].append( ScheduledNode( type=send_recv + cat_str, chunk=_chunk_ if _cat_ == 0 else 1 - _chunk_, stage=stage_, minibatch=_micro_, start_time=complete_time, completion_time=complete_time, ) ) comm_id[local_order[stage_][-1]] = comm_id_counter if _chunk_ == 1 and i > 0: communicate("SEND_", i) communicate("RECV_", i - 1) if _chunk_ == 0 and i < self.n_stage - 1: communicate("SEND_", i) communicate("RECV_", i + 1) comm_id_counter += 1 for rank in range(self.n_stage): # For nodes with the same timestamp on the same stage, communication will be prioritized. def even_breaker(x: ScheduledNode): # Compute nodes are always delayed. if x.type in ["F", "B", "W"]: return comm_id_counter # For comm nodes, order by their unique comm id return comm_id[x] local_order[rank] = list(sorted(local_order[rank], key=lambda x: (x.start_time, even_breaker(x)))) # If a recv with intersects with previous computation, reorder them so that recv # is executed before computation and hence can be overlapped. for i in range(len(local_order[rank])): if ( i > 0 and local_order[rank][i - 1].type in {"F", "B", "W"} and local_order[rank][i].type.startswith("RECV") and "POST_VALIDATION" not in local_order[rank][i].type and local_order[rank][i].start_time <= local_order[rank][i - 1].completion_time ): local_order[rank][i], local_order[rank][i - 1] = local_order[rank][i - 1], local_order[rank][i] local_order_with_rollback = [[] for _ in range(self.n_stage)] for rank in range(self.n_stage): rollback_comm = set() if rank > 0: for node in local_order[rank - 1]: if node.type == "POST_VALIDATION": break if node.type == "SEND_FORWARD": assert node.chunk == 0 rollback_comm.add(node.minibatch) for node in local_order[rank]: if node.type == "RECV_FORWARD" and node.chunk == 0 and node.minibatch in rollback_comm: rollback = True rollback_comm.remove(node.minibatch) else: rollback = False local_order_with_rollback[rank].append( ScheduledNode( type=node.type, chunk=node.chunk, stage=node.stage, minibatch=node.minibatch, start_time=node.start_time, completion_time=node.completion_time, rollback=rollback, ) ) assert len(rollback_comm) == 0 return local_order_with_rollback
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/colossalai/pipeline/schedule/_utils.py
colossalai/pipeline/schedule/_utils.py
from collections import OrderedDict from typing import Any, List, Optional, Tuple import torch import torch.cuda from packaging.version import Version from torch.nn import Module from torch.utils._pytree import SUPPORTED_NODES, TreeSpec, tree_flatten, tree_map, tree_unflatten # this register are for torch under version 1.13.1, maybe removed in the future def _odict_flatten(d: "OrderedDict[Any, Any]") -> Tuple[List[Any], Any]: return list(d.values()), list(d.keys()) def _odict_unflatten(values: List[Any], context: Any) -> "OrderedDict[Any, Any]": return OrderedDict((key, value) for key, value in zip(context, values)) if Version(torch.__version__) <= Version("1.13.1"): try: from torch.utils._pytree import register_pytree_node as _register_pytree_node except ImportError: from torch.utils._pytree import _register_pytree_node _register_pytree_node(OrderedDict, _odict_flatten, _odict_unflatten) def tree_map_hf(fn: Any, pytree: Any): flat_args, spec = tree_flatten_hf(pytree) return tree_unflatten([fn(i) for i in flat_args], spec) # use this flatten function to handle the ModelingOutput Class instance. def tree_flatten_hf(pytree: Any) -> Tuple[List[Any], TreeSpec]: """Flattens a pytree into a list of values an a TreeSpec that can be used to reconstruct the pytree. """ if isinstance(pytree, OrderedDict): node_type = OrderedDict flatten_fn = SUPPORTED_NODES[node_type].flatten_fn child_pytrees, context = flatten_fn(pytree) # Recursively flatten the children result: List[Any] = [] children_specs: List["TreeSpec"] = [] for child in child_pytrees: flat, child_spec = tree_flatten_hf(child) result += flat children_specs.append(child_spec) return result, TreeSpec(node_type, context, children_specs) else: result, tree_spec = tree_flatten(pytree) return result, tree_spec def to_device(x: Any, device: Optional[torch.device] = None) -> Any: """Move object to device if it is a tensor. Args: x (Any): Object to be moved. device (Optional[torch.device], optional): Target device. Defaults to None. Returns: Any: Moved object. """ if isinstance(x, torch.Tensor): return x.to(device) return x def get_batch_size(batch: Any) -> int: """Get the batch size (size of dimension-0) of the first tensor in the batch. Args: batch (Any): Batch to be inspected. Raises: RuntimeError: If no tensor is found in the batch. Returns: int: Batch size. """ data_list, _ = tree_flatten(batch) for data in data_list: if isinstance(data, torch.Tensor): return data.size(0) raise RuntimeError("No tensor found in the batch") def get_micro_batch(batch: Any, start: int, micro_batch_size: int) -> Any: """Get a micro batch of the original batch. Args: batch (Any): Batch to be sliced. start (int): Start index of the micro batch. micro_batch_size (int): Size of the micro batch. Returns: Any: Target micro batch. """ def _get_tensor_slice(x: Any): if isinstance(x, torch.Tensor): return x[start : start + micro_batch_size] return x return tree_map(_get_tensor_slice, batch) def model_forward(model: Module, data: Any, internal_inputs: Optional[dict]) -> Any: """Call model forward function with data and internal inputs. Args: model (Module): Model to be called. data (Any): Data loaded from data iterator. internal_inputs (Optional[dict]): Data from previous stage. It must be a dict or None if it's the first stage. Returns: Any: Outputs of the model. """ if internal_inputs is None: internal_inputs = {} if isinstance(data, (list, tuple)): return model(*data, **internal_inputs) elif isinstance(data, dict): return model(**data, **internal_inputs) return model(data, **internal_inputs) def retain_grad(x: Any) -> None: """Call retain_grad() on a tensor. Args: x (Any): Object to be called. """ if isinstance(x, torch.Tensor) and x.requires_grad: x.retain_grad() def require_grad(x: Any) -> None: """Call require_grad on a tensor. Args: x (Any): Object to be called. """ if isinstance(x, torch.Tensor) and not x.requires_grad: x.requires_grad_() def detach(x: Any) -> Any: """Call detach() on a tensor. Args: x (Any): Object to be called. Returns: Any: The detached object. """ if isinstance(x, torch.Tensor): return x.detach() return x def clone(x: Any) -> Any: """Call clone() on a tensor. Args: x (Any): Object to be called. Returns: Any: The cloned object. """ if isinstance(x, torch.Tensor): return x.clone() return x def release_tensor_data(x: Any) -> Any: """Call untyped_storage().resize_(0) on a tensor. Use to release tensor.data and keep grad_fn. Args: x (Any): Object to be called. Returns: Any: The deallocate .data object. """ if isinstance(x, torch.Tensor): return x.data.untyped_storage().resize_(0) return x def merge_batch(data: List[Any], batch_size_dim=0) -> Any: """Merge micro batches into a batch. Args: data (List[Any]): A list of micro batches. Returns: Any: Merge batch. """ if len(data) == 0: return flattened_data = [] tree_spec = None for d in data: # elems should be an instance of OrderedDict elems, tree_spec = tree_flatten_hf(d) flattened_data.append(elems) merged_data = [] for elem_batch in zip(*flattened_data): if isinstance(elem_batch[0], torch.Tensor): if len(elem_batch[0].shape) == 0: # set loss to None in pipeline outputs merged_data.append(None) else: merged_data.append(torch.cat(elem_batch, dim=batch_size_dim)) else: merged_data.append(list(elem_batch)) return tree_unflatten(merged_data, tree_spec)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/__init__.py
examples/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/fp8/mnist/main.py
examples/community/fp8/mnist/main.py
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved. # # See LICENSE for license information. import argparse import torch import torch.nn as nn import torch.nn.functional as F import torch.optim as optim from torch.optim.lr_scheduler import StepLR from torchvision import datasets, transforms try: from transformer_engine import pytorch as te HAVE_TE = True except (ImportError, ModuleNotFoundError): HAVE_TE = False class Net(nn.Module): def __init__(self, use_te=False): super(Net, self).__init__() self.conv1 = nn.Conv2d(1, 32, 3, 1) self.conv2 = nn.Conv2d(32, 64, 3, 1) self.dropout1 = nn.Dropout(0.25) self.dropout2 = nn.Dropout(0.5) if use_te: self.fc1 = te.Linear(9216, 128) self.fc2 = te.Linear(128, 16) else: self.fc1 = nn.Linear(9216, 128) self.fc2 = nn.Linear(128, 16) self.fc3 = nn.Linear(16, 10) def forward(self, x): """FWD""" x = self.conv1(x) x = F.relu(x) x = self.conv2(x) x = F.relu(x) x = F.max_pool2d(x, 2) x = self.dropout1(x) x = torch.flatten(x, 1) x = self.fc1(x) x = F.relu(x) x = self.dropout2(x) x = self.fc2(x) x = self.fc3(x) output = F.log_softmax(x, dim=1) return output def train(args, model, device, train_loader, optimizer, epoch, use_fp8): """Training function.""" model.train() for batch_idx, (data, target) in enumerate(train_loader): data, target = data.to(device), target.to(device) optimizer.zero_grad() with te.fp8_autocast(enabled=use_fp8): output = model(data) loss = F.nll_loss(output, target) loss.backward() optimizer.step() if batch_idx % args.log_interval == 0: print( f"Train Epoch: {epoch} " f"[{batch_idx * len(data)}/{len(train_loader.dataset)} " f"({100. * batch_idx / len(train_loader):.0f}%)]\t" f"Loss: {loss.item():.6f}" ) if args.dry_run: break def calibrate(model, device, test_loader): """Calibration function.""" model.eval() with torch.no_grad(): for data, target in test_loader: data, target = data.to(device), target.to(device) with te.fp8_autocast(enabled=False, calibrating=True): model(data) def test(model, device, test_loader, use_fp8): """Testing function.""" model.eval() test_loss = 0 correct = 0 with torch.no_grad(): for data, target in test_loader: data, target = data.to(device), target.to(device) with te.fp8_autocast(enabled=use_fp8): output = model(data) test_loss += F.nll_loss(output, target, reduction="sum").item() # sum up batch loss pred = output.argmax(dim=1, keepdim=True) # get the index of the max log-probability correct += pred.eq(target.view_as(pred)).sum().item() test_loss /= len(test_loader.dataset) print( f"\nTest set: Average loss: {test_loss:.4f}, " f"Accuracy: {correct}/{len(test_loader.dataset)} " f"({100. * correct / len(test_loader.dataset):.0f}%)\n" ) def main(): # Training settings parser = argparse.ArgumentParser(description="PyTorch MNIST Example") parser.add_argument( "--batch-size", type=int, default=64, metavar="N", help="input batch size for training (default: 64)", ) parser.add_argument( "--test-batch-size", type=int, default=1000, metavar="N", help="input batch size for testing (default: 1000)", ) parser.add_argument( "--epochs", type=int, default=14, metavar="N", help="number of epochs to train (default: 14)", ) parser.add_argument( "--lr", type=float, default=1.0, metavar="LR", help="learning rate (default: 1.0)", ) parser.add_argument( "--gamma", type=float, default=0.7, metavar="M", help="Learning rate step gamma (default: 0.7)", ) parser.add_argument( "--dry-run", action="store_true", default=False, help="quickly check a single pass", ) parser.add_argument("--seed", type=int, default=1, metavar="S", help="random seed (default: 1)") parser.add_argument( "--log-interval", type=int, default=10, metavar="N", help="how many batches to wait before logging training status", ) parser.add_argument( "--save-model", action="store_true", default=False, help="For Saving the current Model", ) parser.add_argument( "--use-fp8", action="store_true", default=False, help="Use FP8 for inference and training without recalibration" ) parser.add_argument("--use-fp8-infer", action="store_true", default=False, help="Use FP8 inference only") parser.add_argument("--use-te", action="store_true", default=False, help="Use Transformer Engine") args = parser.parse_args() use_cuda = torch.cuda.is_available() if args.use_te or args.use_fp8 or args.use_fp8_infer: assert HAVE_TE, "TransformerEngine not installed." if args.use_fp8 or args.use_fp8_infer: args.use_te = True if args.use_te: assert use_cuda, "CUDA needed for FP8 execution." if args.use_fp8_infer: assert not args.use_fp8, "fp8-infer path currently only supports calibration from a bfloat checkpoint" torch.manual_seed(args.seed) device = torch.device("cuda" if use_cuda else "cpu") train_kwargs = {"batch_size": args.batch_size} test_kwargs = {"batch_size": args.test_batch_size} if use_cuda: cuda_kwargs = {"num_workers": 1, "pin_memory": True, "shuffle": True} train_kwargs.update(cuda_kwargs) test_kwargs.update(cuda_kwargs) transform = transforms.Compose([transforms.ToTensor(), transforms.Normalize((0.1307,), (0.3081,))]) dataset1 = datasets.MNIST("../data", train=True, download=True, transform=transform) dataset2 = datasets.MNIST("../data", train=False, transform=transform) train_loader = torch.utils.data.DataLoader(dataset1, **train_kwargs) test_loader = torch.utils.data.DataLoader(dataset2, **test_kwargs) model = Net(use_te=args.use_te).to(device) optimizer = optim.Adadelta(model.parameters(), lr=args.lr) scheduler = StepLR(optimizer, step_size=1, gamma=args.gamma) for epoch in range(1, args.epochs + 1): train(args, model, device, train_loader, optimizer, epoch, args.use_fp8) test(model, device, test_loader, args.use_fp8) scheduler.step() if args.use_fp8_infer: calibrate(model, device, test_loader) if args.save_model or args.use_fp8_infer: torch.save(model.state_dict(), "mnist_cnn.pt") print("Eval with reloaded checkpoint : fp8=" + str(args.use_fp8_infer)) weights = torch.load("mnist_cnn.pt") model.load_state_dict(weights) test(model, device, test_loader, args.use_fp8_infer) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/nvidia_bert_dataset_provider.py
examples/community/roberta/pretraining/nvidia_bert_dataset_provider.py
import os import random import time from concurrent.futures import ProcessPoolExecutor import h5py import numpy as np import torch import torch.distributed as dist from bert_dataset_provider import BertDatasetProviderInterface from torch.utils.data import DataLoader, Dataset from torch.utils.data.distributed import DistributedSampler # Workaround because python functions are not picklable class WorkerInitObj(object): def __init__(self, seed): self.seed = seed def __call__(self, id): np.random.seed(seed=self.seed + id) random.seed(self.seed + id) def create_pretraining_dataset( input_file, max_predictions_per_seq, num_workers, train_batch_size, worker_init, data_sampler ): train_data = pretraining_dataset(input_file=input_file, max_predictions_per_seq=max_predictions_per_seq) train_dataloader = DataLoader( train_data, sampler=data_sampler(train_data), batch_size=train_batch_size, num_workers=num_workers, worker_init_fn=worker_init, pin_memory=True, ) return train_dataloader, len(train_data) class pretraining_dataset(Dataset): def __init__(self, input_file, max_predictions_per_seq): self.input_file = input_file self.max_predictions_per_seq = max_predictions_per_seq f = h5py.File(input_file, "r") keys = ["input_ids", "input_mask", "segment_ids", "masked_lm_positions"] self.inputs = [np.asarray(f[key][:]) for key in keys] f.close() def __len__(self): "Denotes the total number of samples" return len(self.inputs[0]) def __getitem__(self, index): [input_ids, input_mask, segment_ids, masked_lm_labels] = [ ( torch.from_numpy(input[index].astype(np.int64)) if indice < 5 else torch.from_numpy(np.asarray(input[index].astype(np.int64))) ) for indice, input in enumerate(self.inputs) ] return [input_ids, input_mask, segment_ids, masked_lm_labels] class NvidiaBertDatasetProvider(BertDatasetProviderInterface): def __init__(self, args, evaluate=False): self.num_workers = args.num_workers self.max_seq_length = args.max_seq_length self.max_predictions_per_seq = args.max_predictions_per_seq self.gradient_accumulation_steps = args.gradient_accumulation_steps if not evaluate: self.train_micro_batch_size_per_gpu = args.train_micro_batch_size_per_gpu else: self.train_micro_batch_size_per_gpu = args.eval_micro_batch_size_per_gpu self.logger = args.logger self.global_rank = dist.get_rank() self.world_size = dist.get_world_size() # Initialize dataset files if not evaluate: self.dataset_files = [ os.path.join(args.data_path_prefix, f) for f in os.listdir(args.data_path_prefix) if os.path.isfile(os.path.join(args.data_path_prefix, f)) and "h5" in f ] else: self.dataset_files = [ os.path.join(args.eval_data_path_prefix, f) for f in os.listdir(args.eval_data_path_prefix) if os.path.isfile(os.path.join(args.eval_data_path_prefix, f)) and "h5" in f ] self.dataset_files.sort() # random.shuffle(self.dataset_files) self.num_files = len(self.dataset_files) # self.data_sampler = RandomSampler self.data_sampler = DistributedSampler self.worker_init = WorkerInitObj(args.seed + args.local_rank) self.dataset_future = None self.pool = ProcessPoolExecutor(1) self.data_file = None self.shuffle = True if self.global_rank == 0: self.logger.info(f"NvidiaBertDatasetProvider - Initialization: num_files = {self.num_files}") def get_shard(self, index): start = time.time() if self.dataset_future is None: self.data_file = self._get_shard_file(index) self.train_dataloader, sample_count = create_pretraining_dataset( input_file=self.data_file, max_predictions_per_seq=self.max_predictions_per_seq, num_workers=self.num_workers, train_batch_size=self.train_micro_batch_size_per_gpu, worker_init=self.worker_init, data_sampler=self.data_sampler, ) else: self.train_dataloader, sample_count = self.dataset_future.result(timeout=None) self.logger.info( f"Data Loading Completed for Pretraining Data from {self.data_file} with {sample_count} samples took {time.time()-start:.2f}s." ) return self.train_dataloader, sample_count def release_shard(self): del self.train_dataloader self.pool.shutdown() def prefetch_shard(self, index): self.data_file = self._get_shard_file(index) self.dataset_future = self.pool.submit( create_pretraining_dataset, self.data_file, self.max_predictions_per_seq, self.num_workers, self.train_micro_batch_size_per_gpu, self.worker_init, self.data_sampler, ) def get_batch(self, batch_iter): return batch_iter def prefetch_batch(self): pass def _get_shard_file(self, shard_index): file_index = self._get_shard_file_index(shard_index, self.global_rank) return self.dataset_files[file_index] def _get_shard_file_index(self, shard_index, global_rank): # if dist.is_initialized() and self.world_size > self.num_files: # remainder = self.world_size % self.num_files # file_index = (shard_index * self.world_size) + global_rank + ( # remainder * shard_index) # else: # file_index = shard_index * self.world_size + global_rank return shard_index % self.num_files def shuffle_dataset(self, epoch): if self.shuffle: # deterministically shuffle based on epoch and seed g = torch.Generator() g.manual_seed(self.epoch) indices = torch.randperm(self.num_files, generator=g).tolist() new_dataset = [self.dataset_files[i] for i in indices] self.dataset_files = new_dataset
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/pretrain_utils.py
examples/community/roberta/pretraining/pretrain_utils.py
import os import sys import torch import transformers from transformers import get_linear_schedule_with_warmup from colossalai.legacy.core import global_context as gpc from colossalai.nn.optimizer import HybridAdam sys.path.append(os.getcwd()) from collections import OrderedDict import torch.nn as nn from model.bert import BertForMaskedLM from model.deberta_v2 import DebertaV2ForMaskedLM __all__ = ["get_model", "get_optimizer", "get_lr_scheduler", "get_dataloader_for_pretraining"] def get_new_state_dict(state_dict, start_index=13): new_state_dict = OrderedDict() for k, v in state_dict.items(): name = k[start_index:] new_state_dict[name] = v return new_state_dict class LMModel(nn.Module): def __init__(self, model, config, args): super().__init__() self.checkpoint = args.checkpoint_activations self.config = config self.model = model if self.checkpoint: self.model.gradient_checkpointing_enable() def forward(self, input_ids, token_type_ids=None, attention_mask=None): # Only return lm_logits return self.model(input_ids=input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask) def get_model(args, logger): if args.mlm == "bert": config = transformers.BertConfig.from_json_file(args.bert_config) model = BertForMaskedLM(config) elif args.mlm == "deberta_v2": config = transformers.DebertaV2Config.from_json_file(args.bert_config) model = DebertaV2ForMaskedLM(config) else: raise Exception("Invalid mlm!") if len(args.load_pretrain_model) > 0: assert os.path.exists(args.load_pretrain_model) # load_checkpoint(args.load_pretrain_model, model, strict=False) m_state_dict = torch.load( args.load_pretrain_model, map_location=torch.device(f"cuda:{torch.cuda.current_device()}") ) # new_state_dict = get_new_state_dict(m_state_dict) model.load_state_dict( m_state_dict, strict=True ) # must insure that every process have identical parameters !!!!!!! logger.info("load model success") numel = sum([p.numel() for p in model.parameters()]) if args.checkpoint_activations: model.gradient_checkpointing_enable() # model = LMModel(model, config, args) return config, model, numel def get_optimizer(model, lr): param_optimizer = list(model.named_parameters()) no_decay = ["bias", "gamma", "beta", "LayerNorm"] # configure the weight decay for bert models optimizer_grouped_parameters = [ {"params": [p for n, p in param_optimizer if not any(nd in n for nd in no_decay)], "weight_decay": 0.1}, {"params": [p for n, p in param_optimizer if any(nd in n for nd in no_decay)], "weight_decay": 0.0}, ] optimizer = HybridAdam(optimizer_grouped_parameters, lr=lr, betas=[0.9, 0.95]) return optimizer def get_lr_scheduler(optimizer, total_steps, warmup_steps=2000, last_epoch=-1): # warmup_steps = int(total_steps * warmup_ratio) lr_scheduler = get_linear_schedule_with_warmup( optimizer, num_warmup_steps=warmup_steps, num_training_steps=total_steps, last_epoch=last_epoch ) # lr_scheduler = LinearWarmupLR(optimizer, total_steps=total_steps, warmup_steps=warmup_steps) return lr_scheduler def save_ckpt(model, optimizer, lr_scheduler, path, epoch, shard, global_step): model_path = path + "_pytorch_model.bin" optimizer_lr_path = path + ".op_lrs" checkpoint = {} checkpoint["optimizer"] = optimizer.state_dict() checkpoint["lr_scheduler"] = lr_scheduler.state_dict() checkpoint["epoch"] = epoch checkpoint["shard"] = shard checkpoint["global_step"] = global_step model_state = model.state_dict() # each process must run model.state_dict() if gpc.get_global_rank() == 0: torch.save(checkpoint, optimizer_lr_path) torch.save(model_state, model_path)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/arguments.py
examples/community/roberta/pretraining/arguments.py
import argparse __all__ = ["parse_args"] def parse_args(): parser = argparse.ArgumentParser() parser.add_argument( "--distplan", type=str, default="CAI_Gemini", help="The distributed plan [colossalai, zero1, zero2, torch_ddp, torch_zero].", ) parser.add_argument( "--tp_degree", type=int, default=1, help="Tensor Parallelism Degree. Valid when using colossalai as dist plan.", ) parser.add_argument( "--placement", type=str, default="cpu", help="Placement Policy for Gemini. Valid when using colossalai as dist plan.", ) parser.add_argument( "--shardinit", action="store_true", help="Shard the tensors when init the model to shrink peak memory size on the assigned device. Valid when using colossalai as dist plan.", ) parser.add_argument("--lr", type=float, required=True, help="initial learning rate") parser.add_argument("--epoch", type=int, required=True, help="number of epoch") parser.add_argument("--data_path_prefix", type=str, required=True, help="location of the train data corpus") parser.add_argument( "--eval_data_path_prefix", type=str, required=True, help="location of the evaluation data corpus" ) parser.add_argument("--tokenizer_path", type=str, required=True, help="location of the tokenizer") parser.add_argument("--max_seq_length", type=int, default=512, help="sequence length") parser.add_argument( "--refresh_bucket_size", type=int, default=1, help="This param makes sure that a certain task is repeated for this time steps to \ optimize on the back propagation speed with APEX's DistributedDataParallel", ) parser.add_argument( "--max_predictions_per_seq", "--max_pred", default=80, type=int, help="The maximum number of masked tokens in a sequence to be predicted.", ) parser.add_argument("--gradient_accumulation_steps", default=1, type=int, help="accumulation_steps") parser.add_argument("--train_micro_batch_size_per_gpu", default=2, type=int, required=True, help="train batch size") parser.add_argument("--eval_micro_batch_size_per_gpu", default=2, type=int, required=True, help="eval batch size") parser.add_argument("--num_workers", default=8, type=int, help="") parser.add_argument("--async_worker", action="store_true", help="") parser.add_argument("--bert_config", required=True, type=str, help="location of config.json") parser.add_argument("--wandb", action="store_true", help="use wandb to watch model") parser.add_argument("--wandb_project_name", default="roberta", help="wandb project name") parser.add_argument("--log_interval", default=100, type=int, help="report interval") parser.add_argument("--log_path", type=str, required=True, help="log file which records train step") parser.add_argument("--tensorboard_path", type=str, required=True, help="location of tensorboard file") parser.add_argument( "--colossal_config", type=str, required=True, help="colossal config, which contains zero config and so on" ) parser.add_argument( "--ckpt_path", type=str, required=True, help="location of saving checkpoint, which contains model and optimizer" ) parser.add_argument("--seed", type=int, default=42, help="random seed for initialization") parser.add_argument("--vscode_debug", action="store_true", help="use vscode to debug") parser.add_argument("--load_pretrain_model", default="", type=str, help="location of model's checkpoint") parser.add_argument( "--load_optimizer_lr", default="", type=str, help="location of checkpoint, which contains optimizer, learning rate, epoch, shard and global_step", ) parser.add_argument("--resume_train", action="store_true", help="whether resume training from a early checkpoint") parser.add_argument("--mlm", default="bert", type=str, help="model type, bert or deberta") parser.add_argument("--checkpoint_activations", action="store_true", help="whether to use gradient checkpointing") args = parser.parse_args() return args
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/bert_dataset_provider.py
examples/community/roberta/pretraining/bert_dataset_provider.py
class BertDatasetProviderInterface: def get_shard(self, index, shuffle=True): raise NotImplementedError def release_shard(self, index): raise NotImplementedError def prefetch_shard(self, index): raise NotImplementedError def get_batch(self, batch_iter): raise NotImplementedError def prefetch_batch(self): raise NotImplementedError
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/evaluation.py
examples/community/roberta/pretraining/evaluation.py
import math import os import torch from nvidia_bert_dataset_provider import NvidiaBertDatasetProvider from tqdm import tqdm from utils.global_vars import get_tensorboard_writer, get_timers def evaluate(model, args, logger, global_step, criterion): evaluate_dataset_provider = NvidiaBertDatasetProvider(args, evaluate=True) start_shard = 0 model.eval() timers = get_timers() eval_step = 0 eval_loss = 0 cur_loss = 0 world_size = torch.distributed.get_world_size() with torch.no_grad(): for shard in range(start_shard, len(os.listdir(args.eval_data_path_prefix))): timers("eval_shard_time").start() dataset_iterator, total_length = evaluate_dataset_provider.get_shard(shard) # evaluate_dataset_provider.prefetch_shard(shard + 1) if torch.distributed.get_rank() == 0: iterator_data = tqdm( enumerate(dataset_iterator), total=(total_length // args.eval_micro_batch_size_per_gpu // world_size), colour="MAGENTA", smoothing=1, ) else: iterator_data = enumerate(dataset_iterator) for ( step, batch_data, ) in ( iterator_data ): # tqdm(enumerate(dataset_iterator), total=(total_length // args.train_micro_batch_size_per_gpu // world_size), colour='cyan', smoothing=1): # batch_data = pretrain_dataset_provider.get_batch(batch_index) eval_step += 1 input_ids = batch_data[0].cuda() attention_mask = batch_data[1].cuda() token_type_ids = batch_data[2].cuda() mlm_label = batch_data[3].cuda() # nsp_label = batch_data[5].cuda() output = model(input_ids=input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask) loss = criterion(output.logits, mlm_label) # prediction_scores evaluate_dataset_provider.prefetch_batch() eval_loss += loss.float().item() cur_loss = eval_loss / eval_step elapsed_time = timers("eval_shard_time").elapsed() elapsed_time_per_iteration = elapsed_time / eval_step ppl = math.exp(cur_loss) if args.wandb and torch.distributed.get_rank() == 0: tensorboard_log = get_tensorboard_writer() tensorboard_log.log_eval( {"loss": cur_loss, "ppl": ppl, "mins_batch": elapsed_time_per_iteration}, global_step ) eval_log_str = ( f"evaluation shard: {shard} | step: {eval_step} | elapsed_time: {elapsed_time / 60 :.3f} minutes " + f"| mins/batch: {elapsed_time_per_iteration :.3f} seconds | loss: {cur_loss:.7f} | ppl: {ppl:.7f}" ) logger.info(eval_log_str) logger.info("-" * 100) logger.info("") evaluate_dataset_provider.release_shard() model.train() return cur_loss
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/loss.py
examples/community/roberta/pretraining/loss.py
import torch __all__ = ["LossForPretraining"] class LossForPretraining(torch.nn.Module): def __init__(self, vocab_size): super(LossForPretraining, self).__init__() self.loss_fn = torch.nn.CrossEntropyLoss(ignore_index=-1) self.vocab_size = vocab_size def forward(self, prediction_scores, masked_lm_labels, next_sentence_labels=None): masked_lm_loss = self.loss_fn(prediction_scores.view(-1, self.vocab_size), masked_lm_labels.view(-1)) # next_sentence_loss = self.loss_fn(seq_relationship_score.view(-1, 2), next_sentence_labels.view(-1)) total_loss = masked_lm_loss # + next_sentence_loss return total_loss
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/run_pretraining.py
examples/community/roberta/pretraining/run_pretraining.py
import math import os import time from functools import partial import torch from arguments import parse_args from evaluation import evaluate from loss import LossForPretraining from nvidia_bert_dataset_provider import NvidiaBertDatasetProvider from pretrain_utils import get_lr_scheduler, get_model, get_optimizer, save_ckpt from tqdm import tqdm from transformers import AutoTokenizer from utils.exp_util import get_mem_info, get_tflops, log_args, throughput_calculator from utils.global_vars import get_tensorboard_writer, get_timers, set_global_variables from utils.logger import Logger import colossalai from colossalai.accelerator import get_accelerator from colossalai.context import ParallelMode from colossalai.nn.parallel import zero_model_wrapper, zero_optim_wrapper from colossalai.tensor import ProcessGroup, ShardSpec from colossalai.utils.model.colo_init_context import ColoInitContext def main(): args = parse_args() launch_time = time.strftime("%Y-%m-%d-%H:%M:%S", time.localtime()) tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_path) # os.environ['CUDA_LAUNCH_BLOCKING'] = '1' logger = Logger(os.path.join(args.log_path, launch_time), cuda=torch.cuda.is_available(), debug=args.vscode_debug) if args.vscode_debug: colossalai.launch( rank=args.rank, world_size=args.world_size, host=args.host, port=args.port, backend=args.backend ) args.local_rank = -1 args.log_interval = 1 else: colossalai.launch_from_torch() # args.colossal_config args.local_rank = int(os.environ["LOCAL_RANK"]) logger.info( f"launch_from_torch, world size: {torch.distributed.get_world_size()} | " + f"ParallelMode.MODEL: {ParallelMode.MODEL} | ParallelMode.DATA: {ParallelMode.DATA} | ParallelMode.TENSOR: {ParallelMode.TENSOR}" ) log_args(logger, args) args.tokenizer = tokenizer args.logger = logger set_global_variables(launch_time, args.tensorboard_path) world_size = torch.distributed.get_world_size() get_accelerator().get_current_device() # build model, optimizer and criterion if args.distplan.startswith("CAI"): # all param must use the same process group. world_size = torch.distributed.get_world_size() shard_pg = ProcessGroup(tp_degree=world_size) if args.shardinit else None default_dist_spec = ShardSpec([-1], [world_size]) if args.shardinit else None if args.shardinit and args.distplan != "CAI_Gemini": raise RuntimeError("You can only use shardinit with CAI_Gemini") # build GPT model with ColoInitContext( device=get_accelerator().get_current_device(), dtype=torch.half, default_dist_spec=default_dist_spec, default_pg=shard_pg, ): config, model, numel = get_model(args, logger) # assign running configurations gemini_config = None if args.distplan.startswith("CAI_ZeRO"): optim_config = dict(reduce_bucket_size=12 * 1024 * 1024, overlap_communication=True, verbose=True) elif args.distplan == "CAI_Gemini": gemini_config = dict( strict_ddp_mode=args.tp_degree == 1, device=get_accelerator().get_current_device(), placement_policy=args.placement, pin_memory=True, hidden_dim=model.config.hidden_size, search_range_m=128, ) optim_config = dict(gpu_margin_mem_ratio=0.0) else: raise RuntimeError # build a highly optimized gpu/cpu optimizer optimizer = get_optimizer(model, lr=args.lr) if args.distplan == "CAI_ZeRO1": zero_stage = 1 elif args.distplan == "CAI_ZeRO2": zero_stage = 2 elif args.distplan == "CAI_Gemini": zero_stage = 3 else: raise RuntimeError # wrap your model and optimizer model = zero_model_wrapper(model, zero_stage, gemini_config) optimizer = zero_optim_wrapper(model, optimizer, optim_config=optim_config) logger.info(get_mem_info(prefix="After init optim, ")) else: config, model, numel = get_model(args, logger) logger.info("no_zero") if torch.distributed.get_rank() == 0: os.mkdir(os.path.join(args.ckpt_path, launch_time)) logger.info(f"Model numel: {numel}") get_tflops_func = partial(get_tflops, numel, args.train_micro_batch_size_per_gpu, args.max_seq_length) # 144003367 is is the length of the entire dataset # len(dataloader) steps_per_epoch = ( 144003367 // world_size // args.train_micro_batch_size_per_gpu // args.gradient_accumulation_steps // args.refresh_bucket_size ) total_steps = steps_per_epoch * args.epoch lr_scheduler = get_lr_scheduler(optimizer, total_steps=total_steps, last_epoch=-1) start_epoch = 0 start_shard = 0 global_step = 0 if args.resume_train: assert os.path.exists(args.load_optimizer_lr) o_l_state_dict = torch.load(args.load_optimizer_lr, map_location="cpu") o_l_state_dict["lr_scheduler"]["last_epoch"] = o_l_state_dict["lr_scheduler"]["last_epoch"] - 1 optimizer.load_state_dict(o_l_state_dict["optimizer"]) # o_l_state_dict['lr_scheduler']['last_epoch'] lr_scheduler = get_lr_scheduler( optimizer, total_steps=total_steps, last_epoch=o_l_state_dict["lr_scheduler"]["last_epoch"] ) for state in optimizer.state.values(): for k, v in state.items(): if isinstance(v, torch.Tensor): state[k] = v.cuda(f"cuda:{torch.cuda.current_device()}") # if you want delete the above three code, must move the model to gpu. Because in optimizer.step() lr_scheduler.load_state_dict(o_l_state_dict["lr_scheduler"]) start_epoch = o_l_state_dict["epoch"] start_shard = o_l_state_dict["shard"] + 1 # global_step = o_l_state_dict['global_step'] + 1 logger.info( f"resume from epoch {start_epoch} shard {start_shard} step {lr_scheduler.last_epoch} lr {lr_scheduler.get_last_lr()[0]}" ) criterion = LossForPretraining(config.vocab_size) # build dataloader pretrain_dataset_provider = NvidiaBertDatasetProvider(args) logger.info(get_mem_info(prefix="After init model, ")) eval_loss = 0 train_loss = 0 timers = get_timers() timers("interval_time").start() timers("epoch_time").start() timers("shard_time").start() for epoch in range(start_epoch, args.epoch): for shard in range(start_shard, len(os.listdir(args.data_path_prefix))): dataset_iterator, total_length = pretrain_dataset_provider.get_shard(shard) # pretrain_dataset_provider.prefetch_shard(shard + 1) # may cause cpu memory overload if torch.distributed.get_rank() == 0: iterator_data = tqdm( enumerate(dataset_iterator), total=(total_length // args.train_micro_batch_size_per_gpu // world_size), colour="cyan", smoothing=1, ) else: iterator_data = enumerate(dataset_iterator) model.train() for step, batch_data in iterator_data: # batch_data = pretrain_dataset_provider.get_batch(batch_index) input_ids = batch_data[0].cuda(f"cuda:{torch.cuda.current_device()}") attention_mask = batch_data[1].cuda(f"cuda:{torch.cuda.current_device()}") token_type_ids = batch_data[2].cuda(f"cuda:{torch.cuda.current_device()}") mlm_label = batch_data[3].cuda(f"cuda:{torch.cuda.current_device()}") # nsp_label = batch_data[5].cuda() output = model(input_ids=input_ids, token_type_ids=token_type_ids, attention_mask=attention_mask) loss = criterion(output.logits, mlm_label) pretrain_dataset_provider.prefetch_batch() optimizer.backward(loss) train_loss += loss.float().item() # if (step + 1) % args.accumulation_step == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() global_step += 1 if global_step % args.log_interval == 0 and global_step != 0 and torch.distributed.get_rank() == 0: elapsed_time = timers("interval_time").elapsed(reset=False) elapsed_time_per_iteration = elapsed_time / global_step samples_per_sec, tflops, approx_parameters_in_billions = throughput_calculator( numel, args, config, elapsed_time, global_step, world_size ) cur_loss = train_loss / args.log_interval current_lr = lr_scheduler.get_last_lr()[0] log_str = ( f"| epoch: {epoch} | shard: {shard} | step: {global_step} | lr {current_lr:.7f} | elapsed_time: {elapsed_time / 60 :.3f} minutes " + f"| mins/batch: {elapsed_time_per_iteration :.3f} seconds | loss: {cur_loss:.7f} | ppl: {math.exp(cur_loss):.3f} | TFLOPS: {get_tflops_func(elapsed_time_per_iteration):.3f} or {tflops:.3f}" ) logger.info(log_str, print_=False) if args.wandb: tensorboard_log = get_tensorboard_writer() tensorboard_log.log_train( { "lr": current_lr, "loss": cur_loss, "ppl": math.exp(cur_loss), "mins_batch": elapsed_time_per_iteration, }, global_step, ) train_loss = 0 logger.info(f'epoch {epoch} shard {shard} has cost {timers("shard_time").elapsed() / 60 :.3f} mins') logger.info("*" * 100) eval_loss += evaluate(model, args, logger, global_step, criterion) save_ckpt( model, optimizer, lr_scheduler, os.path.join(args.ckpt_path, launch_time, f"epoch-{epoch}_shard-{shard}_" + launch_time), epoch, shard, global_step, ) eval_loss /= len(os.listdir(args.data_path_prefix)) logger.info( f'epoch {epoch} | shard_length {len(os.listdir(args.data_path_prefix))} | elapsed_time: {timers("epoch_time").elapsed() / 60 :.3f} mins' + f"eval_loss: {eval_loss} | ppl: {math.exp(eval_loss)}" ) logger.info("-" * 100) if args.wandb and torch.distributed.get_rank() == 0: tensorboard_log = get_tensorboard_writer() tensorboard_log.log_eval( { "all_eval_shard_loss": eval_loss, }, epoch, ) start_shard = 0 eval_loss = 0 pretrain_dataset_provider.release_shard() logger.info("Congratulation, training has finished!!!") if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/model/bert.py
examples/community/roberta/pretraining/model/bert.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """PyTorch BERT model.""" import math import os import warnings from dataclasses import dataclass from typing import List, Optional, Tuple, Union import torch import torch.utils.checkpoint from packaging import version from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from transformers.activations import ACT2FN from transformers.modeling_outputs import ( BaseModelOutputWithPastAndCrossAttentions, BaseModelOutputWithPoolingAndCrossAttentions, CausalLMOutputWithCrossAttentions, MaskedLMOutput, MultipleChoiceModelOutput, NextSentencePredictorOutput, QuestionAnsweringModelOutput, SequenceClassifierOutput, TokenClassifierOutput, ) from transformers.modeling_utils import PreTrainedModel from transformers.models.bert.configuration_bert import BertConfig from transformers.pytorch_utils import apply_chunking_to_forward, find_pruneable_heads_and_indices, prune_linear_layer from transformers.utils import ( ModelOutput, add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) logger = logging.get_logger(__name__) _CHECKPOINT_FOR_DOC = "bert-base-uncased" _CONFIG_FOR_DOC = "BertConfig" _TOKENIZER_FOR_DOC = "BertTokenizer" # TokenClassification docstring _CHECKPOINT_FOR_TOKEN_CLASSIFICATION = "dbmdz/bert-large-cased-finetuned-conll03-english" _TOKEN_CLASS_EXPECTED_OUTPUT = ( "['O', 'I-ORG', 'I-ORG', 'I-ORG', 'O', 'O', 'O', 'O', 'O', 'I-LOC', 'O', 'I-LOC', 'I-LOC'] " ) _TOKEN_CLASS_EXPECTED_LOSS = 0.01 # QuestionAnswering docstring _CHECKPOINT_FOR_QA = "deepset/bert-base-cased-squad2" _QA_EXPECTED_OUTPUT = "'a nice puppet'" _QA_EXPECTED_LOSS = 7.41 _QA_TARGET_START_INDEX = 14 _QA_TARGET_END_INDEX = 15 # SequenceClassification docstring _CHECKPOINT_FOR_SEQUENCE_CLASSIFICATION = "textattack/bert-base-uncased-yelp-polarity" _SEQ_CLASS_EXPECTED_OUTPUT = "'LABEL_1'" _SEQ_CLASS_EXPECTED_LOSS = 0.01 BERT_PRETRAINED_MODEL_ARCHIVE_LIST = [ "bert-base-uncased", "bert-large-uncased", "bert-base-cased", "bert-large-cased", "bert-base-multilingual-uncased", "bert-base-multilingual-cased", "bert-base-chinese", "bert-base-german-cased", "bert-large-uncased-whole-word-masking", "bert-large-cased-whole-word-masking", "bert-large-uncased-whole-word-masking-finetuned-squad", "bert-large-cased-whole-word-masking-finetuned-squad", "bert-base-cased-finetuned-mrpc", "bert-base-german-dbmdz-cased", "bert-base-german-dbmdz-uncased", "cl-tohoku/bert-base-japanese", "cl-tohoku/bert-base-japanese-whole-word-masking", "cl-tohoku/bert-base-japanese-char", "cl-tohoku/bert-base-japanese-char-whole-word-masking", "TurkuNLP/bert-base-finnish-cased-v1", "TurkuNLP/bert-base-finnish-uncased-v1", "wietsedv/bert-base-dutch-cased", # See all BERT models at https://huggingface.co/models?filter=bert ] def load_tf_weights_in_bert(model, config, tf_checkpoint_path): """Load tf checkpoints in a pytorch model.""" try: import re import numpy as np import tensorflow as tf except ImportError: logger.error( "Loading a TensorFlow model in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions." ) raise tf_path = os.path.abspath(tf_checkpoint_path) logger.info(f"Converting TensorFlow checkpoint from {tf_path}") # Load weights from TF model init_vars = tf.train.list_variables(tf_path) names = [] arrays = [] for name, shape in init_vars: logger.info(f"Loading TF weight {name} with shape {shape}") array = tf.train.load_variable(tf_path, name) names.append(name) arrays.append(array) for name, array in zip(names, arrays): name = name.split("/") # adam_v and adam_m are variables used in AdamWeightDecayOptimizer to calculated m and v # which are not required for using pretrained model if any( n in ["adam_v", "adam_m", "AdamWeightDecayOptimizer", "AdamWeightDecayOptimizer_1", "global_step"] for n in name ): logger.info(f"Skipping {'/'.join(name)}") continue pointer = model for m_name in name: if re.fullmatch(r"[A-Za-z]+_\d+", m_name): scope_names = re.split(r"_(\d+)", m_name) else: scope_names = [m_name] if scope_names[0] == "kernel" or scope_names[0] == "gamma": pointer = getattr(pointer, "weight") elif scope_names[0] == "output_bias" or scope_names[0] == "beta": pointer = getattr(pointer, "bias") elif scope_names[0] == "output_weights": pointer = getattr(pointer, "weight") elif scope_names[0] == "squad": pointer = getattr(pointer, "classifier") else: try: pointer = getattr(pointer, scope_names[0]) except AttributeError: logger.info(f"Skipping {'/'.join(name)}") continue if len(scope_names) >= 2: num = int(scope_names[1]) pointer = pointer[num] if m_name[-11:] == "_embeddings": pointer = getattr(pointer, "weight") elif m_name == "kernel": array = np.transpose(array) try: if pointer.shape != array.shape: raise ValueError(f"Pointer shape {pointer.shape} and array shape {array.shape} mismatched") except AssertionError as e: e.args += (pointer.shape, array.shape) raise logger.info(f"Initialize PyTorch weight {name}") pointer.data = torch.from_numpy(array) return model class BertEmbeddings(nn.Module): """Construct the embeddings from word, position and token_type embeddings.""" def __init__(self, config): super().__init__() self.word_embeddings = nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) self.position_embeddings = nn.Embedding(config.max_position_embeddings, config.hidden_size) self.token_type_embeddings = nn.Embedding(config.type_vocab_size, config.hidden_size) # self.LayerNorm is not snake-cased to stick with TensorFlow model variable name and be able to load # any TensorFlow checkpoint file self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) self.dropout = nn.Dropout(config.hidden_dropout_prob) # position_ids (1, len position emb) is contiguous in memory and exported when serialized self.position_embedding_type = getattr(config, "position_embedding_type", "absolute") self.register_buffer("position_ids", torch.arange(config.max_position_embeddings).expand((1, -1))) if version.parse(torch.__version__) > version.parse("1.6.0"): self.register_buffer( "token_type_ids", torch.zeros(self.position_ids.size(), dtype=torch.long), persistent=False, ) def forward( self, input_ids: Optional[torch.LongTensor] = None, token_type_ids: Optional[torch.LongTensor] = None, position_ids: Optional[torch.LongTensor] = None, inputs_embeds: Optional[torch.FloatTensor] = None, past_key_values_length: int = 0, ) -> torch.Tensor: if input_ids is not None: input_shape = input_ids.size() else: input_shape = inputs_embeds.size()[:-1] seq_length = input_shape[1] if position_ids is None: position_ids = self.position_ids[:, past_key_values_length : seq_length + past_key_values_length] # Setting the token_type_ids to the registered buffer in constructor where it is all zeros, which usually occurs # when its auto-generated, registered buffer helps users when tracing the model without passing token_type_ids, solves # issue #5664 if token_type_ids is None: if hasattr(self, "token_type_ids"): buffered_token_type_ids = self.token_type_ids[:, :seq_length] buffered_token_type_ids_expanded = buffered_token_type_ids.expand(input_shape[0], seq_length) token_type_ids = buffered_token_type_ids_expanded else: token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=self.position_ids.device) if inputs_embeds is None: inputs_embeds = self.word_embeddings(input_ids) token_type_embeddings = self.token_type_embeddings(token_type_ids) embeddings = inputs_embeds + token_type_embeddings if self.position_embedding_type == "absolute": position_embeddings = self.position_embeddings(position_ids) embeddings += position_embeddings embeddings = self.LayerNorm(embeddings) embeddings = self.dropout(embeddings) return embeddings class BertSelfAttention(nn.Module): def __init__(self, config, position_embedding_type=None): super().__init__() if config.hidden_size % config.num_attention_heads != 0 and not hasattr(config, "embedding_size"): raise ValueError( f"The hidden size ({config.hidden_size}) is not a multiple of the number of attention " f"heads ({config.num_attention_heads})" ) self.num_attention_heads = config.num_attention_heads self.attention_head_size = int(config.hidden_size / config.num_attention_heads) self.all_head_size = self.num_attention_heads * self.attention_head_size self.query = nn.Linear(config.hidden_size, self.all_head_size) self.key = nn.Linear(config.hidden_size, self.all_head_size) self.value = nn.Linear(config.hidden_size, self.all_head_size) self.dropout = nn.Dropout(config.attention_probs_dropout_prob) self.position_embedding_type = position_embedding_type or getattr(config, "position_embedding_type", "absolute") if self.position_embedding_type == "relative_key" or self.position_embedding_type == "relative_key_query": self.max_position_embeddings = config.max_position_embeddings self.distance_embedding = nn.Embedding(2 * config.max_position_embeddings - 1, self.attention_head_size) self.is_decoder = config.is_decoder def transpose_for_scores(self, x: torch.Tensor) -> torch.Tensor: new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) x = x.view(new_x_shape) return x.permute(0, 2, 1, 3) def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.FloatTensor] = None, head_mask: Optional[torch.FloatTensor] = None, encoder_hidden_states: Optional[torch.FloatTensor] = None, encoder_attention_mask: Optional[torch.FloatTensor] = None, past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, output_attentions: Optional[bool] = False, ) -> Tuple[torch.Tensor]: mixed_query_layer = self.query(hidden_states) # If this is instantiated as a cross-attention module, the keys # and values come from an encoder; the attention mask needs to be # such that the encoder's padding tokens are not attended to. is_cross_attention = encoder_hidden_states is not None if is_cross_attention and past_key_value is not None: # reuse k,v, cross_attentions key_layer = past_key_value[0] value_layer = past_key_value[1] attention_mask = encoder_attention_mask elif is_cross_attention: key_layer = self.transpose_for_scores(self.key(encoder_hidden_states)) value_layer = self.transpose_for_scores(self.value(encoder_hidden_states)) attention_mask = encoder_attention_mask elif past_key_value is not None: key_layer = self.transpose_for_scores(self.key(hidden_states)) value_layer = self.transpose_for_scores(self.value(hidden_states)) key_layer = torch.cat([past_key_value[0], key_layer], dim=2) value_layer = torch.cat([past_key_value[1], value_layer], dim=2) else: key_layer = self.transpose_for_scores(self.key(hidden_states)) value_layer = self.transpose_for_scores(self.value(hidden_states)) query_layer = self.transpose_for_scores(mixed_query_layer) if self.is_decoder: # if cross_attention save Tuple(torch.Tensor, torch.Tensor) of all cross attention key/value_states. # Further calls to cross_attention layer can then reuse all cross-attention # key/value_states (first "if" case) # if uni-directional self-attention (decoder) save Tuple(torch.Tensor, torch.Tensor) of # all previous decoder key/value_states. Further calls to uni-directional self-attention # can concat previous decoder key/value_states to current projected key/value_states (third "elif" case) # if encoder bi-directional self-attention `past_key_value` is always `None` past_key_value = (key_layer, value_layer) # Take the dot product between "query" and "key" to get the raw attention scores. attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) if self.position_embedding_type == "relative_key" or self.position_embedding_type == "relative_key_query": seq_length = hidden_states.size()[1] position_ids_l = torch.arange(seq_length, dtype=torch.long, device=hidden_states.device).view(-1, 1) position_ids_r = torch.arange(seq_length, dtype=torch.long, device=hidden_states.device).view(1, -1) distance = position_ids_l - position_ids_r positional_embedding = self.distance_embedding(distance + self.max_position_embeddings - 1) positional_embedding = positional_embedding.to(dtype=query_layer.dtype) # fp16 compatibility if self.position_embedding_type == "relative_key": relative_position_scores = torch.einsum("bhld,lrd->bhlr", query_layer, positional_embedding) attention_scores = attention_scores + relative_position_scores elif self.position_embedding_type == "relative_key_query": relative_position_scores_query = torch.einsum("bhld,lrd->bhlr", query_layer, positional_embedding) relative_position_scores_key = torch.einsum("bhld,lrd->bhlr", key_layer, positional_embedding) attention_scores = attention_scores + relative_position_scores_query + relative_position_scores_key attention_scores = attention_scores / math.sqrt(self.attention_head_size) if attention_mask is not None: # Apply the attention mask is (precomputed for all layers in BertModel forward() function) attention_scores = attention_scores + attention_mask # Normalize the attention scores to probabilities. attention_probs = nn.functional.softmax(attention_scores, dim=-1) # This is actually dropping out entire tokens to attend to, which might # seem a bit unusual, but is taken from the original Transformer paper. attention_probs = self.dropout(attention_probs) # Mask heads if we want to if head_mask is not None: attention_probs = attention_probs * head_mask context_layer = torch.matmul(attention_probs, value_layer) context_layer = context_layer.permute(0, 2, 1, 3).contiguous() new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) context_layer = context_layer.view(new_context_layer_shape) outputs = (context_layer, attention_probs) if output_attentions else (context_layer,) if self.is_decoder: outputs = outputs + (past_key_value,) return outputs class BertSelfOutput(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) hidden_states = self.LayerNorm(hidden_states + input_tensor) return hidden_states class BertAttention(nn.Module): def __init__(self, config, position_embedding_type=None): super().__init__() self.self = BertSelfAttention(config, position_embedding_type=position_embedding_type) self.output = BertSelfOutput(config) self.pruned_heads = set() def prune_heads(self, heads): if len(heads) == 0: return heads, index = find_pruneable_heads_and_indices( heads, self.self.num_attention_heads, self.self.attention_head_size, self.pruned_heads ) # Prune linear layers self.self.query = prune_linear_layer(self.self.query, index) self.self.key = prune_linear_layer(self.self.key, index) self.self.value = prune_linear_layer(self.self.value, index) self.output.dense = prune_linear_layer(self.output.dense, index, dim=1) # Update hyper params and store pruned heads self.self.num_attention_heads = self.self.num_attention_heads - len(heads) self.self.all_head_size = self.self.attention_head_size * self.self.num_attention_heads self.pruned_heads = self.pruned_heads.union(heads) def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.FloatTensor] = None, head_mask: Optional[torch.FloatTensor] = None, encoder_hidden_states: Optional[torch.FloatTensor] = None, encoder_attention_mask: Optional[torch.FloatTensor] = None, past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, output_attentions: Optional[bool] = False, ) -> Tuple[torch.Tensor]: self_outputs = self.self( hidden_states, attention_mask, head_mask, encoder_hidden_states, encoder_attention_mask, past_key_value, output_attentions, ) attention_output = self.output(self_outputs[0], hidden_states) outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them return outputs class BertIntermediate(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.intermediate_size) if isinstance(config.hidden_act, str): self.intermediate_act_fn = ACT2FN[config.hidden_act] else: self.intermediate_act_fn = config.hidden_act def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.intermediate_act_fn(hidden_states) return hidden_states class BertOutput(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.intermediate_size, config.hidden_size) self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) hidden_states = self.LayerNorm(hidden_states + input_tensor) return hidden_states class BertLayer(nn.Module): def __init__(self, config): super().__init__() self.chunk_size_feed_forward = config.chunk_size_feed_forward self.seq_len_dim = 1 self.attention = BertAttention(config) self.is_decoder = config.is_decoder self.add_cross_attention = config.add_cross_attention if self.add_cross_attention: if not self.is_decoder: raise ValueError(f"{self} should be used as a decoder model if cross attention is added") self.crossattention = BertAttention(config, position_embedding_type="absolute") self.intermediate = BertIntermediate(config) self.output = BertOutput(config) def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.FloatTensor] = None, head_mask: Optional[torch.FloatTensor] = None, encoder_hidden_states: Optional[torch.FloatTensor] = None, encoder_attention_mask: Optional[torch.FloatTensor] = None, past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, output_attentions: Optional[bool] = False, ) -> Tuple[torch.Tensor]: # decoder uni-directional self-attention cached key/values tuple is at positions 1,2 self_attn_past_key_value = past_key_value[:2] if past_key_value is not None else None self_attention_outputs = self.attention( hidden_states, attention_mask, head_mask, output_attentions=output_attentions, past_key_value=self_attn_past_key_value, ) attention_output = self_attention_outputs[0] # if decoder, the last output is tuple of self-attn cache if self.is_decoder: outputs = self_attention_outputs[1:-1] present_key_value = self_attention_outputs[-1] else: outputs = self_attention_outputs[1:] # add self attentions if we output attention weights cross_attn_present_key_value = None if self.is_decoder and encoder_hidden_states is not None: if not hasattr(self, "crossattention"): raise ValueError( f"If `encoder_hidden_states` are passed, {self} has to be instantiated with cross-attention layers" " by setting `config.add_cross_attention=True`" ) # cross_attn cached key/values tuple is at positions 3,4 of past_key_value tuple cross_attn_past_key_value = past_key_value[-2:] if past_key_value is not None else None cross_attention_outputs = self.crossattention( attention_output, attention_mask, head_mask, encoder_hidden_states, encoder_attention_mask, cross_attn_past_key_value, output_attentions, ) attention_output = cross_attention_outputs[0] outputs = outputs + cross_attention_outputs[1:-1] # add cross attentions if we output attention weights # add cross-attn cache to positions 3,4 of present_key_value tuple cross_attn_present_key_value = cross_attention_outputs[-1] present_key_value = present_key_value + cross_attn_present_key_value layer_output = apply_chunking_to_forward( self.feed_forward_chunk, self.chunk_size_feed_forward, self.seq_len_dim, attention_output ) outputs = (layer_output,) + outputs # if decoder, return the attn key/values as the last output if self.is_decoder: outputs = outputs + (present_key_value,) return outputs def feed_forward_chunk(self, attention_output): intermediate_output = self.intermediate(attention_output) layer_output = self.output(intermediate_output, attention_output) return layer_output class BertEncoder(nn.Module): def __init__(self, config): super().__init__() self.config = config self.layer = nn.ModuleList([BertLayer(config) for _ in range(config.num_hidden_layers)]) self.gradient_checkpointing = False def forward( self, hidden_states: torch.Tensor, attention_mask: Optional[torch.FloatTensor] = None, head_mask: Optional[torch.FloatTensor] = None, encoder_hidden_states: Optional[torch.FloatTensor] = None, encoder_attention_mask: Optional[torch.FloatTensor] = None, past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, use_cache: Optional[bool] = None, output_attentions: Optional[bool] = False, output_hidden_states: Optional[bool] = False, return_dict: Optional[bool] = True, ) -> Union[Tuple[torch.Tensor], BaseModelOutputWithPastAndCrossAttentions]: all_hidden_states = () if output_hidden_states else None all_self_attentions = () if output_attentions else None all_cross_attentions = () if output_attentions and self.config.add_cross_attention else None next_decoder_cache = () if use_cache else None for i, layer_module in enumerate(self.layer): if output_hidden_states: all_hidden_states = all_hidden_states + (hidden_states,) layer_head_mask = head_mask[i] if head_mask is not None else None past_key_value = past_key_values[i] if past_key_values is not None else None if self.gradient_checkpointing and self.training: if use_cache: logger.warning( "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." ) use_cache = False def create_custom_forward(module): def custom_forward(*inputs): return module(*inputs, past_key_value, output_attentions) return custom_forward layer_outputs = torch.utils.checkpoint.checkpoint( create_custom_forward(layer_module), hidden_states, attention_mask, layer_head_mask, encoder_hidden_states, encoder_attention_mask, ) else: layer_outputs = layer_module( hidden_states, attention_mask, layer_head_mask, encoder_hidden_states, encoder_attention_mask, past_key_value, output_attentions, ) hidden_states = layer_outputs[0] if use_cache: next_decoder_cache += (layer_outputs[-1],) if output_attentions: all_self_attentions = all_self_attentions + (layer_outputs[1],) if self.config.add_cross_attention: all_cross_attentions = all_cross_attentions + (layer_outputs[2],) if output_hidden_states: all_hidden_states = all_hidden_states + (hidden_states,) if not return_dict: return tuple( v for v in [ hidden_states, next_decoder_cache, all_hidden_states, all_self_attentions, all_cross_attentions, ] if v is not None ) return BaseModelOutputWithPastAndCrossAttentions( last_hidden_state=hidden_states, past_key_values=next_decoder_cache, hidden_states=all_hidden_states, attentions=all_self_attentions, cross_attentions=all_cross_attentions, ) class BertPooler(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.activation = nn.Tanh() def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: # We "pool" the model by simply taking the hidden state corresponding # to the first token. first_token_tensor = hidden_states[:, 0] pooled_output = self.dense(first_token_tensor) pooled_output = self.activation(pooled_output) return pooled_output class BertPredictionHeadTransform(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) if isinstance(config.hidden_act, str): self.transform_act_fn = ACT2FN[config.hidden_act] else: self.transform_act_fn = config.hidden_act self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.transform_act_fn(hidden_states) hidden_states = self.LayerNorm(hidden_states) return hidden_states class BertLMPredictionHead(nn.Module): def __init__(self, config): super().__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False) self.bias = nn.Parameter(torch.zeros(config.vocab_size)) # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings` self.decoder.bias = self.bias def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) return hidden_states class BertOnlyMLMHead(nn.Module): def __init__(self, config): super().__init__() self.predictions = BertLMPredictionHead(config) def forward(self, sequence_output: torch.Tensor) -> torch.Tensor: prediction_scores = self.predictions(sequence_output) return prediction_scores class BertOnlyNSPHead(nn.Module): def __init__(self, config): super().__init__() self.seq_relationship = nn.Linear(config.hidden_size, 2) def forward(self, pooled_output): seq_relationship_score = self.seq_relationship(pooled_output) return seq_relationship_score class BertPreTrainingHeads(nn.Module): def __init__(self, config): super().__init__() self.predictions = BertLMPredictionHead(config) self.seq_relationship = nn.Linear(config.hidden_size, 2) def forward(self, sequence_output, pooled_output): prediction_scores = self.predictions(sequence_output) seq_relationship_score = self.seq_relationship(pooled_output) return prediction_scores, seq_relationship_score class BertPreTrainedModel(PreTrainedModel): """ An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained models. """ config_class = BertConfig load_tf_weights = load_tf_weights_in_bert base_model_prefix = "bert" supports_gradient_checkpointing = True _keys_to_ignore_on_load_missing = [r"position_ids"] def _init_weights(self, module): """Initialize the weights""" if isinstance(module, nn.Linear): # Slightly different from the TF version which uses truncated_normal for initialization # cf https://github.com/pytorch/pytorch/pull/5617 module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) if module.bias is not None: module.bias.data.zero_() elif isinstance(module, nn.Embedding): module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) if module.padding_idx is not None:
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/model/deberta_v2.py
examples/community/roberta/pretraining/model/deberta_v2.py
# coding=utf-8 # Copyright 2020 Microsoft and the Hugging Face Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ PyTorch DeBERTa-v2 model.""" import math from collections.abc import Sequence from typing import Optional, Tuple, Union import numpy as np import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, LayerNorm, MSELoss from transformers.activations import ACT2FN from transformers.modeling_outputs import ( BaseModelOutput, MaskedLMOutput, MultipleChoiceModelOutput, QuestionAnsweringModelOutput, SequenceClassifierOutput, TokenClassifierOutput, ) from transformers.modeling_utils import PreTrainedModel from transformers.models.deberta_v2.configuration_deberta_v2 import DebertaV2Config from transformers.pytorch_utils import softmax_backward_data from transformers.utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, ) logger = logging.get_logger(__name__) _CONFIG_FOR_DOC = "DebertaV2Config" _TOKENIZER_FOR_DOC = "DebertaV2Tokenizer" _CHECKPOINT_FOR_DOC = "microsoft/deberta-v2-xlarge" DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST = [ "microsoft/deberta-v2-xlarge", "microsoft/deberta-v2-xxlarge", "microsoft/deberta-v2-xlarge-mnli", "microsoft/deberta-v2-xxlarge-mnli", ] # Copied from transformers.models.deberta.modeling_deberta.ContextPooler class ContextPooler(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.pooler_hidden_size, config.pooler_hidden_size) self.dropout = StableDropout(config.pooler_dropout) self.config = config def forward(self, hidden_states): # We "pool" the model by simply taking the hidden state corresponding # to the first token. context_token = hidden_states[:, 0] context_token = self.dropout(context_token) pooled_output = self.dense(context_token) pooled_output = ACT2FN[self.config.pooler_hidden_act](pooled_output) return pooled_output @property def output_dim(self): return self.config.hidden_size # Copied from transformers.models.deberta.modeling_deberta.XSoftmax with deberta->deberta_v2 class XSoftmax(torch.autograd.Function): """ Masked Softmax which is optimized for saving memory Args: input (`torch.tensor`): The input tensor that will apply softmax. mask (`torch.IntTensor`): The mask matrix where 0 indicate that element will be ignored in the softmax calculation. dim (int): The dimension that will apply softmax Example: ```python >>> import torch >>> from transformers.models.deberta_v2.modeling_deberta_v2 import XSoftmax >>> # Make a tensor >>> x = torch.randn([4, 20, 100]) >>> # Create a mask >>> mask = (x > 0).int() >>> # Specify the dimension to apply softmax >>> dim = -1 >>> y = XSoftmax.apply(x, mask, dim) ```""" @staticmethod def forward(self, input, mask, dim): self.dim = dim rmask = ~(mask.to(torch.bool)) output = input.masked_fill(rmask, torch.tensor(torch.finfo(input.dtype).min)) output = torch.softmax(output, self.dim) output.masked_fill_(rmask, 0) self.save_for_backward(output) return output @staticmethod def backward(self, grad_output): (output,) = self.saved_tensors inputGrad = softmax_backward_data(self, grad_output, output, self.dim, output) return inputGrad, None, None @staticmethod def symbolic(g, self, mask, dim): import torch.onnx.symbolic_helper as sym_help from torch.onnx.symbolic_opset9 import masked_fill, softmax mask_cast_value = g.op("Cast", mask, to_i=sym_help.cast_pytorch_to_onnx["Long"]) r_mask = g.op( "Cast", g.op("Sub", g.op("Constant", value_t=torch.tensor(1, dtype=torch.int64)), mask_cast_value), to_i=sym_help.cast_pytorch_to_onnx["Byte"], ) output = masked_fill( g, self, r_mask, g.op("Constant", value_t=torch.tensor(torch.finfo(self.type().dtype()).min)) ) output = softmax(g, output, dim) return masked_fill(g, output, r_mask, g.op("Constant", value_t=torch.tensor(0, dtype=torch.uint8))) # Copied from transformers.models.deberta.modeling_deberta.DropoutContext class DropoutContext(object): def __init__(self): self.dropout = 0 self.mask = None self.scale = 1 self.reuse_mask = True # Copied from transformers.models.deberta.modeling_deberta.get_mask def get_mask(input, local_context): if not isinstance(local_context, DropoutContext): dropout = local_context mask = None else: dropout = local_context.dropout dropout *= local_context.scale mask = local_context.mask if local_context.reuse_mask else None if dropout > 0 and mask is None: mask = (1 - torch.empty_like(input).bernoulli_(1 - dropout)).to(torch.bool) if isinstance(local_context, DropoutContext): if local_context.mask is None: local_context.mask = mask return mask, dropout # Copied from transformers.models.deberta.modeling_deberta.XDropout class XDropout(torch.autograd.Function): """Optimized dropout function to save computation and memory by using mask operation instead of multiplication.""" @staticmethod def forward(ctx, input, local_ctx): mask, dropout = get_mask(input, local_ctx) ctx.scale = 1.0 / (1 - dropout) if dropout > 0: ctx.save_for_backward(mask) return input.masked_fill(mask, 0) * ctx.scale else: return input @staticmethod def backward(ctx, grad_output): if ctx.scale > 1: (mask,) = ctx.saved_tensors return grad_output.masked_fill(mask, 0) * ctx.scale, None else: return grad_output, None # Copied from transformers.models.deberta.modeling_deberta.StableDropout class StableDropout(nn.Module): """ Optimized dropout module for stabilizing the training Args: drop_prob (float): the dropout probabilities """ def __init__(self, drop_prob): super().__init__() self.drop_prob = drop_prob self.count = 0 self.context_stack = None def forward(self, x): """ Call the module Args: x (`torch.tensor`): The input tensor to apply dropout """ if self.training and self.drop_prob > 0: return XDropout.apply(x, self.get_context()) return x def clear_context(self): self.count = 0 self.context_stack = None def init_context(self, reuse_mask=True, scale=1): if self.context_stack is None: self.context_stack = [] self.count = 0 for c in self.context_stack: c.reuse_mask = reuse_mask c.scale = scale def get_context(self): if self.context_stack is not None: if self.count >= len(self.context_stack): self.context_stack.append(DropoutContext()) ctx = self.context_stack[self.count] ctx.dropout = self.drop_prob self.count += 1 return ctx else: return self.drop_prob # Copied from transformers.models.deberta.modeling_deberta.DebertaSelfOutput with DebertaLayerNorm->LayerNorm class DebertaV2SelfOutput(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.LayerNorm = LayerNorm(config.hidden_size, config.layer_norm_eps) self.dropout = StableDropout(config.hidden_dropout_prob) def forward(self, hidden_states, input_tensor): hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) hidden_states = self.LayerNorm(hidden_states + input_tensor) return hidden_states # Copied from transformers.models.deberta.modeling_deberta.DebertaAttention with Deberta->DebertaV2 class DebertaV2Attention(nn.Module): def __init__(self, config): super().__init__() self.self = DisentangledSelfAttention(config) self.output = DebertaV2SelfOutput(config) self.config = config def forward( self, hidden_states, attention_mask, output_attentions=False, query_states=None, relative_pos=None, rel_embeddings=None, ): self_output = self.self( hidden_states, attention_mask, output_attentions, query_states=query_states, relative_pos=relative_pos, rel_embeddings=rel_embeddings, ) if output_attentions: self_output, att_matrix = self_output if query_states is None: query_states = hidden_states attention_output = self.output(self_output, query_states) if output_attentions: return (attention_output, att_matrix) else: return attention_output # Copied from transformers.models.bert.modeling_bert.BertIntermediate with Bert->DebertaV2 class DebertaV2Intermediate(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.hidden_size, config.intermediate_size) if isinstance(config.hidden_act, str): self.intermediate_act_fn = ACT2FN[config.hidden_act] else: self.intermediate_act_fn = config.hidden_act def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: hidden_states = self.dense(hidden_states) hidden_states = self.intermediate_act_fn(hidden_states) return hidden_states # Copied from transformers.models.deberta.modeling_deberta.DebertaOutput with DebertaLayerNorm->LayerNorm class DebertaV2Output(nn.Module): def __init__(self, config): super().__init__() self.dense = nn.Linear(config.intermediate_size, config.hidden_size) self.LayerNorm = LayerNorm(config.hidden_size, config.layer_norm_eps) self.dropout = StableDropout(config.hidden_dropout_prob) self.config = config def forward(self, hidden_states, input_tensor): hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) hidden_states = self.LayerNorm(hidden_states + input_tensor) return hidden_states # Copied from transformers.models.deberta.modeling_deberta.DebertaLayer with Deberta->DebertaV2 class DebertaV2Layer(nn.Module): def __init__(self, config): super().__init__() self.attention = DebertaV2Attention(config) self.intermediate = DebertaV2Intermediate(config) self.output = DebertaV2Output(config) def forward( self, hidden_states, attention_mask, query_states=None, relative_pos=None, rel_embeddings=None, output_attentions=False, ): attention_output = self.attention( hidden_states, attention_mask, output_attentions=output_attentions, query_states=query_states, relative_pos=relative_pos, rel_embeddings=rel_embeddings, ) if output_attentions: attention_output, att_matrix = attention_output intermediate_output = self.intermediate(attention_output) layer_output = self.output(intermediate_output, attention_output) if output_attentions: return (layer_output, att_matrix) else: return layer_output class ConvLayer(nn.Module): def __init__(self, config): super().__init__() kernel_size = getattr(config, "conv_kernel_size", 3) groups = getattr(config, "conv_groups", 1) self.conv_act = getattr(config, "conv_act", "tanh") self.conv = nn.Conv1d( config.hidden_size, config.hidden_size, kernel_size, padding=(kernel_size - 1) // 2, groups=groups ) self.LayerNorm = LayerNorm(config.hidden_size, config.layer_norm_eps) self.dropout = StableDropout(config.hidden_dropout_prob) self.config = config def forward(self, hidden_states, residual_states, input_mask): out = self.conv(hidden_states.permute(0, 2, 1).contiguous()).permute(0, 2, 1).contiguous() rmask = (1 - input_mask).bool() out.masked_fill_(rmask.unsqueeze(-1).expand(out.size()), 0) out = ACT2FN[self.conv_act](self.dropout(out)) layer_norm_input = residual_states + out output = self.LayerNorm(layer_norm_input).to(layer_norm_input) if input_mask is None: output_states = output else: if input_mask.dim() != layer_norm_input.dim(): if input_mask.dim() == 4: input_mask = input_mask.squeeze(1).squeeze(1) input_mask = input_mask.unsqueeze(2) input_mask = input_mask.to(output.dtype) output_states = output * input_mask return output_states class DebertaV2Encoder(nn.Module): """Modified BertEncoder with relative position bias support""" def __init__(self, config): super().__init__() self.layer = nn.ModuleList([DebertaV2Layer(config) for _ in range(config.num_hidden_layers)]) self.relative_attention = getattr(config, "relative_attention", False) if self.relative_attention: self.max_relative_positions = getattr(config, "max_relative_positions", -1) if self.max_relative_positions < 1: self.max_relative_positions = config.max_position_embeddings self.position_buckets = getattr(config, "position_buckets", -1) pos_ebd_size = self.max_relative_positions * 2 if self.position_buckets > 0: pos_ebd_size = self.position_buckets * 2 # rel = nn.Parameter(torch.empty((pos_ebd_size, config.hidden_size))) # self.rel_embeddings = nn.init.normal_(rel, mean=0.0, std=config.initializer_range) self.rel_embeddings = nn.Embedding(pos_ebd_size, config.hidden_size) self.norm_rel_ebd = [x.strip() for x in getattr(config, "norm_rel_ebd", "none").lower().split("|")] if "layer_norm" in self.norm_rel_ebd: self.LayerNorm = LayerNorm(config.hidden_size, config.layer_norm_eps, elementwise_affine=True) self.conv = ConvLayer(config) if getattr(config, "conv_kernel_size", 0) > 0 else None self.gradient_checkpointing = False def get_rel_embedding(self): att_span = self.position_buckets rel_index = torch.arange(0, att_span * 2).long().to(self.rel_embeddings.weight.device) rel_embeddings = self.rel_embeddings(rel_index) # rel_embeddings = self.rel_embeddings.weight if self.relative_attention else None # rel_embeddings = self.rel_embeddings if self.relative_attention else None if rel_embeddings is not None and ("layer_norm" in self.norm_rel_ebd): rel_embeddings = self.LayerNorm(rel_embeddings) return rel_embeddings def get_attention_mask(self, attention_mask): if attention_mask.dim() <= 2: extended_attention_mask = attention_mask.unsqueeze(1).unsqueeze(2) attention_mask = extended_attention_mask * extended_attention_mask.squeeze(-2).unsqueeze(-1) attention_mask = attention_mask.byte() elif attention_mask.dim() == 3: attention_mask = attention_mask.unsqueeze(1) return attention_mask def get_rel_pos(self, hidden_states, query_states=None, relative_pos=None): if self.relative_attention and relative_pos is None: q = query_states.size(-2) if query_states is not None else hidden_states.size(-2) relative_pos = build_relative_position( q, hidden_states.size(-2), bucket_size=self.position_buckets, max_position=self.max_relative_positions ) return relative_pos def forward( self, hidden_states, attention_mask, output_hidden_states=True, output_attentions=False, query_states=None, relative_pos=None, return_dict=True, ): if attention_mask.dim() <= 2: input_mask = attention_mask else: input_mask = (attention_mask.sum(-2) > 0).byte() attention_mask = self.get_attention_mask(attention_mask) relative_pos = self.get_rel_pos(hidden_states, query_states, relative_pos) all_hidden_states = () if output_hidden_states else None all_attentions = () if output_attentions else None if isinstance(hidden_states, Sequence): next_kv = hidden_states[0] else: next_kv = hidden_states rel_embeddings = self.get_rel_embedding() output_states = next_kv for i, layer_module in enumerate(self.layer): if output_hidden_states: all_hidden_states = all_hidden_states + (output_states,) if self.gradient_checkpointing and self.training: def create_custom_forward(module): def custom_forward(*inputs): return module(*inputs, output_attentions) return custom_forward output_states = torch.utils.checkpoint.checkpoint( create_custom_forward(layer_module), next_kv, attention_mask, query_states, relative_pos, rel_embeddings, ) else: output_states = layer_module( next_kv, attention_mask, query_states=query_states, relative_pos=relative_pos, rel_embeddings=rel_embeddings, output_attentions=output_attentions, ) if output_attentions: output_states, att_m = output_states if i == 0 and self.conv is not None: output_states = self.conv(hidden_states, output_states, input_mask) if query_states is not None: query_states = output_states if isinstance(hidden_states, Sequence): next_kv = hidden_states[i + 1] if i + 1 < len(self.layer) else None else: next_kv = output_states if output_attentions: all_attentions = all_attentions + (att_m,) if output_hidden_states: all_hidden_states = all_hidden_states + (output_states,) if not return_dict: return tuple(v for v in [output_states, all_hidden_states, all_attentions] if v is not None) return BaseModelOutput( last_hidden_state=output_states, hidden_states=all_hidden_states, attentions=all_attentions ) def make_log_bucket_position(relative_pos, bucket_size, max_position): sign = np.sign(relative_pos) mid = bucket_size // 2 abs_pos = np.where((relative_pos < mid) & (relative_pos > -mid), mid - 1, np.abs(relative_pos)) log_pos = np.ceil(np.log(abs_pos / mid) / np.log((max_position - 1) / mid) * (mid - 1)) + mid bucket_pos = np.where(abs_pos <= mid, relative_pos, log_pos * sign).astype(np.int) return bucket_pos def build_relative_position(query_size, key_size, bucket_size=-1, max_position=-1): """ Build relative position according to the query and key We assume the absolute position of query \\(P_q\\) is range from (0, query_size) and the absolute position of key \\(P_k\\) is range from (0, key_size), The relative positions from query to key is \\(R_{q \\rightarrow k} = P_q - P_k\\) Args: query_size (int): the length of query key_size (int): the length of key bucket_size (int): the size of position bucket max_position (int): the maximum allowed absolute position Return: `torch.LongTensor`: A tensor with shape [1, query_size, key_size] """ q_ids = np.arange(0, query_size) k_ids = np.arange(0, key_size) rel_pos_ids = q_ids[:, None] - np.tile(k_ids, (q_ids.shape[0], 1)) if bucket_size > 0 and max_position > 0: rel_pos_ids = make_log_bucket_position(rel_pos_ids, bucket_size, max_position) rel_pos_ids = torch.tensor(rel_pos_ids, dtype=torch.long) rel_pos_ids = rel_pos_ids[:query_size, :] rel_pos_ids = rel_pos_ids.unsqueeze(0) return rel_pos_ids @torch.jit.script # Copied from transformers.models.deberta.modeling_deberta.c2p_dynamic_expand def c2p_dynamic_expand(c2p_pos, query_layer, relative_pos): return c2p_pos.expand([query_layer.size(0), query_layer.size(1), query_layer.size(2), relative_pos.size(-1)]) @torch.jit.script # Copied from transformers.models.deberta.modeling_deberta.p2c_dynamic_expand def p2c_dynamic_expand(c2p_pos, query_layer, key_layer): return c2p_pos.expand([query_layer.size(0), query_layer.size(1), key_layer.size(-2), key_layer.size(-2)]) @torch.jit.script # Copied from transformers.models.deberta.modeling_deberta.pos_dynamic_expand def pos_dynamic_expand(pos_index, p2c_att, key_layer): return pos_index.expand(p2c_att.size()[:2] + (pos_index.size(-2), key_layer.size(-2))) class DisentangledSelfAttention(nn.Module): """ Disentangled self-attention module Parameters: config (`DebertaV2Config`): A model config class instance with the configuration to build a new model. The schema is similar to *BertConfig*, for more details, please refer [`DebertaV2Config`] """ def __init__(self, config): super().__init__() if config.hidden_size % config.num_attention_heads != 0: raise ValueError( f"The hidden size ({config.hidden_size}) is not a multiple of the number of attention " f"heads ({config.num_attention_heads})" ) self.num_attention_heads = config.num_attention_heads _attention_head_size = config.hidden_size // config.num_attention_heads self.attention_head_size = getattr(config, "attention_head_size", _attention_head_size) self.all_head_size = self.num_attention_heads * self.attention_head_size self.query_proj = nn.Linear(config.hidden_size, self.all_head_size, bias=True) self.key_proj = nn.Linear(config.hidden_size, self.all_head_size, bias=True) self.value_proj = nn.Linear(config.hidden_size, self.all_head_size, bias=True) self.share_att_key = getattr(config, "share_att_key", False) self.pos_att_type = config.pos_att_type if config.pos_att_type is not None else [] self.relative_attention = getattr(config, "relative_attention", False) if self.relative_attention: self.position_buckets = getattr(config, "position_buckets", -1) self.max_relative_positions = getattr(config, "max_relative_positions", -1) if self.max_relative_positions < 1: self.max_relative_positions = config.max_position_embeddings self.pos_ebd_size = self.max_relative_positions if self.position_buckets > 0: self.pos_ebd_size = self.position_buckets self.pos_dropout = StableDropout(config.hidden_dropout_prob) if not self.share_att_key: if "c2p" in self.pos_att_type: self.pos_key_proj = nn.Linear(config.hidden_size, self.all_head_size, bias=True) if "p2c" in self.pos_att_type: self.pos_query_proj = nn.Linear(config.hidden_size, self.all_head_size) self.dropout = StableDropout(config.attention_probs_dropout_prob) # self.LayerNorm = LayerNorm(config.hidden_size, config.layer_norm_eps, elementwise_affine=True) def transpose_for_scores(self, x, attention_heads): new_x_shape = x.size()[:-1] + (attention_heads, -1) x = x.view(new_x_shape) return x.permute(0, 2, 1, 3).contiguous().view(-1, x.size(1), x.size(-1)) def forward( self, hidden_states, attention_mask, output_attentions=False, query_states=None, relative_pos=None, rel_embeddings=None, ): """ Call the module Args: hidden_states (`torch.FloatTensor`): Input states to the module usually the output from previous layer, it will be the Q,K and V in *Attention(Q,K,V)* attention_mask (`torch.ByteTensor`): An attention mask matrix of shape [*B*, *N*, *N*] where *B* is the batch size, *N* is the maximum sequence length in which element [i,j] = *1* means the *i* th token in the input can attend to the *j* th token. output_attentions (`bool`, optional): Whether return the attention matrix. query_states (`torch.FloatTensor`, optional): The *Q* state in *Attention(Q,K,V)*. relative_pos (`torch.LongTensor`): The relative position encoding between the tokens in the sequence. It's of shape [*B*, *N*, *N*] with values ranging in [*-max_relative_positions*, *max_relative_positions*]. rel_embeddings (`torch.FloatTensor`): The embedding of relative distances. It's a tensor of shape [\\(2 \\times \\text{max_relative_positions}\\), *hidden_size*]. """ if query_states is None: query_states = hidden_states query_layer = self.transpose_for_scores(self.query_proj(query_states), self.num_attention_heads) key_layer = self.transpose_for_scores(self.key_proj(hidden_states), self.num_attention_heads) value_layer = self.transpose_for_scores(self.value_proj(hidden_states), self.num_attention_heads) rel_att = None # Take the dot product between "query" and "key" to get the raw attention scores. scale_factor = 1 if "c2p" in self.pos_att_type: scale_factor += 1 if "p2c" in self.pos_att_type: scale_factor += 1 scale = math.sqrt(query_layer.size(-1) * scale_factor) attention_scores = torch.bmm(query_layer, key_layer.transpose(-1, -2)) / scale if self.relative_attention: rel_embeddings = self.pos_dropout(rel_embeddings) rel_att = self.disentangled_attention_bias( query_layer, key_layer, relative_pos, rel_embeddings, scale_factor ) if rel_att is not None: attention_scores = attention_scores + rel_att attention_scores = attention_scores attention_scores = attention_scores.view( -1, self.num_attention_heads, attention_scores.size(-2), attention_scores.size(-1) ) # bsz x height x length x dimension attention_probs = XSoftmax.apply(attention_scores, attention_mask, -1) attention_probs = self.dropout(attention_probs) context_layer = torch.bmm( attention_probs.view(-1, attention_probs.size(-2), attention_probs.size(-1)), value_layer ) context_layer = ( context_layer.view(-1, self.num_attention_heads, context_layer.size(-2), context_layer.size(-1)) .permute(0, 2, 1, 3) .contiguous() ) new_context_layer_shape = context_layer.size()[:-2] + (-1,) context_layer = context_layer.view(new_context_layer_shape) if output_attentions: return (context_layer, attention_probs) else: return context_layer def disentangled_attention_bias(self, query_layer, key_layer, relative_pos, rel_embeddings, scale_factor): if relative_pos is None: q = query_layer.size(-2) relative_pos = build_relative_position( q, key_layer.size(-2), bucket_size=self.position_buckets, max_position=self.max_relative_positions ) if relative_pos.dim() == 2: relative_pos = relative_pos.unsqueeze(0).unsqueeze(0) elif relative_pos.dim() == 3: relative_pos = relative_pos.unsqueeze(1) # bsz x height x query x key elif relative_pos.dim() != 4: raise ValueError(f"Relative position ids must be of dim 2 or 3 or 4. {relative_pos.dim()}") att_span = self.pos_ebd_size relative_pos = relative_pos.long().to(query_layer.device) # rel_index = torch.arange(0, att_span * 2).long().to(query_layer.device) # rel_embeddings = rel_embeddings(rel_index).unsqueeze(0) rel_embeddings = rel_embeddings.unsqueeze(0) # rel_embeddings = rel_embeddings.unsqueeze(0) # rel_embeddings = rel_embeddings[0 : att_span * 2, :].unsqueeze(0) if self.share_att_key: pos_query_layer = self.transpose_for_scores( self.query_proj(rel_embeddings), self.num_attention_heads ).repeat(query_layer.size(0) // self.num_attention_heads, 1, 1) pos_key_layer = self.transpose_for_scores(self.key_proj(rel_embeddings), self.num_attention_heads).repeat( query_layer.size(0) // self.num_attention_heads, 1, 1 ) else: if "c2p" in self.pos_att_type: pos_key_layer = self.transpose_for_scores( self.pos_key_proj(rel_embeddings), self.num_attention_heads ).repeat( query_layer.size(0) // self.num_attention_heads, 1, 1 ) # .split(self.all_head_size, dim=-1) if "p2c" in self.pos_att_type: pos_query_layer = self.transpose_for_scores( self.pos_query_proj(rel_embeddings), self.num_attention_heads ).repeat( query_layer.size(0) // self.num_attention_heads, 1, 1 ) # .split(self.all_head_size, dim=-1) score = 0 # content->position if "c2p" in self.pos_att_type: scale = math.sqrt(pos_key_layer.size(-1) * scale_factor) c2p_att = torch.bmm(query_layer, pos_key_layer.transpose(-1, -2)) c2p_pos = torch.clamp(relative_pos + att_span, 0, att_span * 2 - 1) c2p_att = torch.gather( c2p_att, dim=-1, index=c2p_pos.squeeze(0).expand([query_layer.size(0), query_layer.size(1), relative_pos.size(-1)]), ) score += c2p_att / scale # position->content if "p2c" in self.pos_att_type: scale = math.sqrt(pos_query_layer.size(-1) * scale_factor) if key_layer.size(-2) != query_layer.size(-2): r_pos = build_relative_position( key_layer.size(-2), key_layer.size(-2), bucket_size=self.position_buckets, max_position=self.max_relative_positions, ).to(query_layer.device) r_pos = r_pos.unsqueeze(0) else: r_pos = relative_pos p2c_pos = torch.clamp(-r_pos + att_span, 0, att_span * 2 - 1) p2c_att = torch.bmm(key_layer, pos_query_layer.transpose(-1, -2)) p2c_att = torch.gather( p2c_att, dim=-1, index=p2c_pos.squeeze(0).expand([query_layer.size(0), key_layer.size(-2), key_layer.size(-2)]), ).transpose(-1, -2) score += p2c_att / scale return score # Copied from transformers.models.deberta.modeling_deberta.DebertaEmbeddings with DebertaLayerNorm->LayerNorm class DebertaV2Embeddings(nn.Module): """Construct the embeddings from word, position and token_type embeddings.""" def __init__(self, config): super().__init__() pad_token_id = getattr(config, "pad_token_id", 0)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
true
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/global_vars.py
examples/community/roberta/pretraining/utils/global_vars.py
import time import torch from .WandbLog import TensorboardLog _GLOBAL_TIMERS = None _GLOBAL_TENSORBOARD_WRITER = None def set_global_variables(launch_time, tensorboard_path): _set_timers() _set_tensorboard_writer(launch_time, tensorboard_path) def _set_timers(): """Initialize timers.""" global _GLOBAL_TIMERS _ensure_var_is_not_initialized(_GLOBAL_TIMERS, "timers") _GLOBAL_TIMERS = Timers() def _set_tensorboard_writer(launch_time, tensorboard_path): """Set tensorboard writer.""" global _GLOBAL_TENSORBOARD_WRITER _ensure_var_is_not_initialized(_GLOBAL_TENSORBOARD_WRITER, "tensorboard writer") if torch.distributed.get_rank() == 0: _GLOBAL_TENSORBOARD_WRITER = TensorboardLog(tensorboard_path + f"/{launch_time}", launch_time) def get_timers(): """Return timers.""" _ensure_var_is_initialized(_GLOBAL_TIMERS, "timers") return _GLOBAL_TIMERS def get_tensorboard_writer(): """Return tensorboard writer. It can be None so no need to check if it is initialized.""" return _GLOBAL_TENSORBOARD_WRITER def _ensure_var_is_initialized(var, name): """Make sure the input variable is not None.""" assert var is not None, "{} is not initialized.".format(name) def _ensure_var_is_not_initialized(var, name): """Make sure the input variable is not None.""" assert var is None, "{} is already initialized.".format(name) class _Timer: """Timer.""" def __init__(self, name): self.name_ = name self.elapsed_ = 0.0 self.started_ = False self.start_time = time.time() def start(self): """Start the timer.""" # assert not self.started_, 'timer has already been started' torch.cuda.synchronize() self.start_time = time.time() self.started_ = True def stop(self): """Stop the timer.""" assert self.started_, "timer is not started" torch.cuda.synchronize() self.elapsed_ += time.time() - self.start_time self.started_ = False def reset(self): """Reset timer.""" self.elapsed_ = 0.0 self.started_ = False def elapsed(self, reset=True): """Calculate the elapsed time.""" started_ = self.started_ # If the timing in progress, end it first. if self.started_: self.stop() # Get the elapsed time. elapsed_ = self.elapsed_ # Reset the elapsed time if reset: self.reset() # If timing was in progress, set it back. if started_: self.start() return elapsed_ class Timers: """Group of timers.""" def __init__(self): self.timers = {} def __call__(self, name): if name not in self.timers: self.timers[name] = _Timer(name) return self.timers[name] def write(self, names, writer, iteration, normalizer=1.0, reset=False): """Write timers to a tensorboard writer""" # currently when using add_scalars, # torch.utils.add_scalars makes each timer its own run, which # pollutes the runs list, so we just add each as a scalar assert normalizer > 0.0 for name in names: value = self.timers[name].elapsed(reset=reset) / normalizer writer.add_scalar(name + "-time", value, iteration) def log(self, names, normalizer=1.0, reset=True): """Log a group of timers.""" assert normalizer > 0.0 string = "time (ms)" for name in names: elapsed_time = self.timers[name].elapsed(reset=reset) * 1000.0 / normalizer string += " | {}: {:.2f}".format(name, elapsed_time) if torch.distributed.is_initialized(): if torch.distributed.get_rank() == (torch.distributed.get_world_size() - 1): print(string, flush=True) else: print(string, flush=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/logger.py
examples/community/roberta/pretraining/utils/logger.py
import logging import torch.distributed as dist logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s", datefmt="%m/%d/%Y %H:%M:%S", level=logging.INFO ) logger = logging.getLogger(__name__) class Logger: def __init__(self, log_path, cuda=False, debug=False): self.logger = logging.getLogger(__name__) self.cuda = cuda self.log_path = log_path self.debug = debug def info(self, message, log_=True, print_=True, *args, **kwargs): if (self.cuda and dist.get_rank() == 0) or not self.cuda: if print_: self.logger.info(message, *args, **kwargs) if log_: with open(self.log_path, "a+") as f_log: f_log.write(message + "\n") def error(self, message, *args, **kwargs): self.logger.error(message, *args, **kwargs)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/exp_util.py
examples/community/roberta/pretraining/utils/exp_util.py
import functools import os import shutil import psutil import torch from colossalai.legacy.core import global_context as gpc def logging(s, log_path, print_=True, log_=True): if print_: print(s) if log_: with open(log_path, "a+") as f_log: f_log.write(s + "\n") def get_logger(log_path, **kwargs): return functools.partial(logging, log_path=log_path, **kwargs) def create_exp_dir(dir_path, scripts_to_save=None, debug=False): if debug: print("Debug Mode : no experiment dir created") return functools.partial(logging, log_path=None, log_=False) if not os.path.exists(dir_path): os.makedirs(dir_path) print("Experiment dir : {}".format(dir_path)) if scripts_to_save is not None: script_path = os.path.join(dir_path, "scripts") if not os.path.exists(script_path): os.makedirs(script_path) for script in scripts_to_save: dst_file = os.path.join(dir_path, "scripts", os.path.basename(script)) shutil.copyfile(script, dst_file) return get_logger(log_path=os.path.join(dir_path, "log.txt")) def get_cpu_mem(): return psutil.Process().memory_info().rss / 1024**2 def get_gpu_mem(): return torch.cuda.memory_allocated() / 1024**2 def get_mem_info(prefix=""): return f"{prefix}GPU memory usage: {get_gpu_mem():.2f} MB, CPU memory usage: {get_cpu_mem():.2f} MB" def get_tflops(model_numel, batch_size, seq_len, step_time): return model_numel * batch_size * seq_len * 8 / 1e12 / (step_time + 1e-12) def get_parameters_in_billions(model, world_size=1): gpus_per_model = world_size approx_parameters_in_billions = sum( [ sum([p.ds_numel if hasattr(p, "ds_id") else p.nelement() for p in model_module.parameters()]) for model_module in model ] ) return approx_parameters_in_billions * gpus_per_model / (1e9) def throughput_calculator(numel, args, config, iteration_time, total_iterations, world_size=1): gpus_per_model = 1 batch_size = args.train_micro_batch_size_per_gpu batch_size * args.max_seq_length world_size / gpus_per_model approx_parameters_in_billions = numel elapsed_time_per_iter = iteration_time / total_iterations samples_per_second = batch_size / elapsed_time_per_iter # flops calculator hidden_size = config.hidden_size num_layers = config.num_hidden_layers vocab_size = config.vocab_size # General TFLOPs formula (borrowed from Equation 3 in Section 5.1 of # https://arxiv.org/pdf/2104.04473.pdf). # The factor of 4 is when used with activation check-pointing, # otherwise it will be 3. checkpoint_activations_factor = 4 if args.checkpoint_activations else 3 flops_per_iteration = ( 24 * checkpoint_activations_factor * batch_size * args.max_seq_length * num_layers * (hidden_size**2) ) * (1.0 + (args.max_seq_length / (6.0 * hidden_size)) + (vocab_size / (16.0 * num_layers * hidden_size))) tflops = flops_per_iteration / (elapsed_time_per_iter * (10**12)) return samples_per_second, tflops, approx_parameters_in_billions def synchronize(): if not torch.distributed.is_available(): return if not torch.distributed.is_initialized(): return world_size = torch.distributed.get_world_size() if world_size == 1: return torch.distributed.barrier() def log_args(logger, args): logger.info("--------args----------") message = "\n".join([f"{k:<30}: {v}" for k, v in vars(args).items()]) message += "\n" message += "\n".join([f"{k:<30}: {v}" for k, v in gpc.config.items()]) logger.info(message) logger.info("--------args----------\n")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/pretraining/utils/WandbLog.py
examples/community/roberta/pretraining/utils/WandbLog.py
import os import time import wandb from torch.utils.tensorboard import SummaryWriter class WandbLog: @classmethod def init_wandb(cls, project, notes=None, name=time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()), config=None): wandb.init(project=project, notes=notes, name=name, config=config) @classmethod def log(cls, result, model=None, gradient=None): wandb.log(result) if model: wandb.watch(model) if gradient: wandb.watch(gradient) class TensorboardLog: def __init__(self, location, name=time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()), config=None): if not os.path.exists(location): os.mkdir(location) self.writer = SummaryWriter(location, comment=name) def log_train(self, result, step): for k, v in result.items(): self.writer.add_scalar(f"{k}/train", v, step) def log_eval(self, result, step): for k, v in result.items(): self.writer.add_scalar(f"{k}/eval", v, step) def log_zeroshot(self, result, step): for k, v in result.items(): self.writer.add_scalar(f"{k}_acc/eval", v, step)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/tokenize_mask.py
examples/community/roberta/preprocessing/tokenize_mask.py
import argparse import multiprocessing import os import time from random import shuffle import h5py import numpy as np import psutil from get_mask import PreTrainingDataset from tqdm import tqdm from transformers import AutoTokenizer def get_raw_instance(document, max_sequence_length=512): """ Get the initial training instances, split the whole segment into multiple parts according to the max_sequence_length, and return as multiple processed instances. :param document: document :param max_sequence_length: :return: a list. each element is a sequence of text """ # document = self.documents[index] max_sequence_length_allowed = max_sequence_length - 2 # document = [seq for seq in document if len(seq)<max_sequence_length_allowed] sizes = [len(seq) for seq in document] result_list = [] curr_seq = [] sz_idx = 0 while sz_idx < len(sizes): if len(curr_seq) + sizes[sz_idx] <= max_sequence_length_allowed: # or len(curr_seq)==0: curr_seq += document[sz_idx] sz_idx += 1 elif sizes[sz_idx] >= max_sequence_length_allowed: if len(curr_seq) > 0: result_list.append(curr_seq) curr_seq = [] result_list.append(document[sz_idx][:max_sequence_length_allowed]) sz_idx += 1 else: result_list.append(curr_seq) curr_seq = [] if len(curr_seq) > max_sequence_length_allowed / 2: # /2 result_list.append(curr_seq) # num_instance=int(len(big_list)/max_sequence_length_allowed)+1 # print("num_instance:",num_instance) # result_list=[] # for j in range(num_instance): # index=j*max_sequence_length_allowed # end_index=index+max_sequence_length_allowed if j!=num_instance-1 else -1 # result_list.append(big_list[index:end_index]) return result_list def split_numpy_chunk(path, tokenizer, pretrain_data, host): documents = [] instances = [] s = time.time() with open(path, encoding="utf-8") as fd: document = [] for i, line in enumerate(tqdm(fd)): line = line.strip() # document = line # if len(document.split("<sep>")) <= 3: # continue if len(line) > 0 and line[:2] == "]]": # This is end of document documents.append(document) document = [] elif len(line) >= 2: document.append(line) if len(document) > 0: documents.append(document) print("read_file ", time.time() - s) # documents = [x for x in documents if x] # print(len(documents)) # print(len(documents[0])) # print(documents[0][0:10]) ans = [] for docs in tqdm(documents): ans.append(pretrain_data.tokenize(docs)) print(time.time() - s) del documents instances = [] for a in tqdm(ans): raw_ins = get_raw_instance(a) instances.extend(raw_ins) del ans print("len instance", len(instances)) sen_num = len(instances) seq_len = 512 input_ids = np.zeros([sen_num, seq_len], dtype=np.int32) input_mask = np.zeros([sen_num, seq_len], dtype=np.int32) segment_ids = np.zeros([sen_num, seq_len], dtype=np.int32) masked_lm_output = np.zeros([sen_num, seq_len], dtype=np.int32) for index, ins in tqdm(enumerate(instances)): mask_dict = pretrain_data.create_training_instance(ins) input_ids[index] = mask_dict[0] input_mask[index] = mask_dict[1] segment_ids[index] = mask_dict[2] masked_lm_output[index] = mask_dict[3] with h5py.File(f"/output/{host}.h5", "w") as hf: hf.create_dataset("input_ids", data=input_ids) hf.create_dataset("input_mask", data=input_ids) hf.create_dataset("segment_ids", data=segment_ids) hf.create_dataset("masked_lm_positions", data=masked_lm_output) del instances def split_numpy_chunk_pool(input_path, output_path, pretrain_data, worker, dupe_factor, seq_len, file_name): if os.path.exists(os.path.join(output_path, f"{file_name}.h5")): print(f"{file_name}.h5 exists") return documents = [] instances = [] s = time.time() with open(input_path, "r", encoding="utf-8") as fd: document = [] for i, line in enumerate(tqdm(fd)): line = line.strip() if len(line) > 0 and line[:2] == "]]": # This is end of document documents.append(document) document = [] elif len(line) >= 2: document.append(line) if len(document) > 0: documents.append(document) print(f"read_file cost {time.time() - s}, length is {len(documents)}") ans = [] s = time.time() pool = multiprocessing.Pool(worker) encoded_doc = pool.imap_unordered(pretrain_data.tokenize, documents, 100) for index, res in tqdm(enumerate(encoded_doc, start=1), total=len(documents), colour="cyan"): ans.append(res) pool.close() print((time.time() - s) / 60) del documents instances = [] for a in tqdm(ans, colour="MAGENTA"): raw_ins = get_raw_instance(a, max_sequence_length=seq_len) instances.extend(raw_ins) del ans print("len instance", len(instances)) new_instances = [] for _ in range(dupe_factor): for ins in instances: new_instances.append(ins) shuffle(new_instances) instances = new_instances print("after dupe_factor, len instance", len(instances)) sentence_num = len(instances) input_ids = np.zeros([sentence_num, seq_len], dtype=np.int32) input_mask = np.zeros([sentence_num, seq_len], dtype=np.int32) segment_ids = np.zeros([sentence_num, seq_len], dtype=np.int32) masked_lm_output = np.zeros([sentence_num, seq_len], dtype=np.int32) s = time.time() pool = multiprocessing.Pool(worker) encoded_docs = pool.imap_unordered(pretrain_data.create_training_instance, instances, 32) for index, mask_dict in tqdm(enumerate(encoded_docs), total=len(instances), colour="blue"): input_ids[index] = mask_dict[0] input_mask[index] = mask_dict[1] segment_ids[index] = mask_dict[2] masked_lm_output[index] = mask_dict[3] pool.close() print((time.time() - s) / 60) with h5py.File(os.path.join(output_path, f"{file_name}.h5"), "w") as hf: hf.create_dataset("input_ids", data=input_ids) hf.create_dataset("input_mask", data=input_mask) hf.create_dataset("segment_ids", data=segment_ids) hf.create_dataset("masked_lm_positions", data=masked_lm_output) del instances if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--tokenizer_path", type=str, required=True, default=10, help="path of tokenizer") parser.add_argument("--seq_len", type=int, default=512, help="sequence length") parser.add_argument( "--max_predictions_per_seq", type=int, default=80, help="number of shards, e.g., 10, 50, or 100" ) parser.add_argument("--input_path", type=str, required=True, help="input path of shard which has split sentence") parser.add_argument("--output_path", type=str, required=True, help="output path of h5 contains token id") parser.add_argument( "--backend", type=str, default="python", help="backend of mask token, python, c++, numpy respectively" ) parser.add_argument( "--dupe_factor", type=int, default=1, help="specifies how many times the preprocessor repeats to create the input from the same article/document", ) parser.add_argument("--worker", type=int, default=32, help="number of process") parser.add_argument("--server_num", type=int, default=10, help="number of servers") args = parser.parse_args() tokenizer = AutoTokenizer.from_pretrained(args.tokenizer_path) pretrain_data = PreTrainingDataset( tokenizer, args.seq_len, args.backend, max_predictions_per_seq=args.max_predictions_per_seq ) data_len = len(os.listdir(args.input_path)) for i in range(data_len): input_path = os.path.join(args.input_path, f"{i}.txt") if os.path.exists(input_path): start = time.time() print(f"process {input_path}") split_numpy_chunk_pool( input_path, args.output_path, pretrain_data, args.worker, args.dupe_factor, args.seq_len, i ) end_ = time.time() print("memory:%.4f GB" % (psutil.Process(os.getpid()).memory_info().rss / 1024 / 1024 / 1024)) print(f"has cost {(end_ - start) / 60}") print("-" * 100) print("") # if you have multiple server, you can use code below or modify code to openmpi # host = int(socket.gethostname().split('GPU')[-1]) # for i in range(data_len // args.server_num + 1): # h = args.server_num * i + host - 1 # input_path = os.path.join(args.input_path, f'{h}.txt') # if os.path.exists(input_path): # start = time.time() # print(f'I am server {host}, process {input_path}') # split_numpy_chunk_pool(input_path, # args.output_path, # pretrain_data, # args.worker, # args.dupe_factor, # args.seq_len, # h) # end_ = time.time() # print(u'memory:%.4f GB' % (psutil.Process(os.getpid()).memory_info().rss / 1024 / 1024 / 1024) ) # print(f'has cost {(end_ - start) / 60}') # print('-' * 100) # print('')
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/get_mask.py
examples/community/roberta/preprocessing/get_mask.py
import collections import logging import random import jieba jieba.setLogLevel(logging.CRITICAL) import re import mask import numpy as np PAD = 0 MaskedLMInstance = collections.namedtuple("MaskedLMInstance", ["index", "label"]) def map_to_numpy(data): return np.asarray(data) class PreTrainingDataset: def __init__( self, tokenizer, max_seq_length, backend="python", max_predictions_per_seq: int = 80, do_whole_word_mask: bool = True, ): self.tokenizer = tokenizer self.max_seq_length = max_seq_length self.masked_lm_prob = 0.15 self.backend = backend self.do_whole_word_mask = do_whole_word_mask self.max_predictions_per_seq = max_predictions_per_seq self.vocab_words = list(tokenizer.vocab.keys()) self.rec = re.compile("[\u4E00-\u9FA5]") self.whole_rec = re.compile("##[\u4E00-\u9FA5]") self.mlm_p = 0.15 self.mlm_mask_p = 0.8 self.mlm_tamper_p = 0.05 self.mlm_maintain_p = 0.1 def tokenize(self, doc): temp = [] for d in doc: temp.append(self.tokenizer.tokenize(d)) return temp def create_training_instance(self, instance): is_next = 1 raw_text_list = self.get_new_segment(instance) tokens_a = raw_text_list assert len(tokens_a) == len(instance) # tokens_a, tokens_b, is_next = instance.get_values() # print(f'is_next label:{is_next}') # Create mapper tokens = [] original_tokens = [] segment_ids = [] tokens.append("[CLS]") original_tokens.append("[CLS]") segment_ids.append(0) for index, token in enumerate(tokens_a): tokens.append(token) original_tokens.append(instance[index]) segment_ids.append(0) tokens.append("[SEP]") original_tokens.append("[SEP]") segment_ids.append(0) # for token in tokens_b: # tokens.append(token) # segment_ids.append(1) # tokens.append("[SEP]") # segment_ids.append(1) # Get Masked LM predictions if self.backend == "c++": output_tokens, masked_lm_output = mask.create_whole_masked_lm_predictions( tokens, original_tokens, self.vocab_words, self.tokenizer.vocab, self.max_predictions_per_seq, self.masked_lm_prob, ) elif self.backend == "python": output_tokens, masked_lm_output = self.create_whole_masked_lm_predictions(tokens) # Convert to Ids input_ids = self.tokenizer.convert_tokens_to_ids(output_tokens) input_mask = [1] * len(input_ids) while len(input_ids) < self.max_seq_length: input_ids.append(PAD) segment_ids.append(PAD) input_mask.append(PAD) masked_lm_output.append(-1) return [ map_to_numpy(input_ids), map_to_numpy(input_mask), map_to_numpy(segment_ids), map_to_numpy(masked_lm_output), map_to_numpy([is_next]), ] def create_masked_lm_predictions(self, tokens): cand_indexes = [] for i, token in enumerate(tokens): if token == "[CLS]" or token == "[SEP]": continue if self.do_whole_word_mask and len(cand_indexes) >= 1 and token.startswith("##"): cand_indexes[-1].append(i) else: cand_indexes.append([i]) # cand_indexes.append(i) random.shuffle(cand_indexes) output_tokens = list(tokens) num_to_predict = min(self.max_predictions_per_seq, max(1, int(round(len(tokens) * self.masked_lm_prob)))) masked_lms = [] covered_indexes = set() for index in cand_indexes: if len(masked_lms) >= num_to_predict: break if index in covered_indexes: continue covered_indexes.add(index) masked_token = None # 80% mask if random.random() < 0.8: masked_token = "[MASK]" else: # 10% Keep Original if random.random() < 0.5: masked_token = tokens[index] # 10% replace w/ random word else: masked_token = self.vocab_words[random.randint(0, len(self.vocab_words) - 1)] output_tokens[index] = masked_token masked_lms.append(MaskedLMInstance(index=index, label=tokens[index])) masked_lms = sorted(masked_lms, key=lambda x: x.index) masked_lm_output = [-1] * len(output_tokens) for p in masked_lms: masked_lm_output[p.index] = self.tokenizer.vocab[p.label] return (output_tokens, masked_lm_output) def get_new_segment(self, segment): """ Input a sentence, return a processed sentence: In order to support the Chinese whole word mask, the words that are separated will be marked with a special mark ("#"), so that the subsequent processing module can know which words belong to the same word. :param segment: a sentence """ seq_cws = jieba.lcut("".join(segment)) seq_cws_dict = {x: 1 for x in seq_cws} new_segment = [] i = 0 while i < len(segment): if len(self.rec.findall(segment[i])) == 0: new_segment.append(segment[i]) i += 1 continue has_add = False for length in range(3, 0, -1): if i + length > len(segment): continue if "".join(segment[i : i + length]) in seq_cws_dict: new_segment.append(segment[i]) for l in range(1, length): new_segment.append("##" + segment[i + l]) i += length has_add = True break if not has_add: new_segment.append(segment[i]) i += 1 return new_segment def create_whole_masked_lm_predictions(self, tokens): """Creates the predictions for the masked LM objective.""" cand_indexes = [] for i, token in enumerate(tokens): if token == "[CLS]" or token == "[SEP]": continue # Whole Word Masking means that if we mask all of the wordpieces # corresponding to an original word. When a word has been split into # WordPieces, the first token does not have any marker and any subsequence # tokens are prefixed with ##. So whenever we see the ## token, we # append it to the previous set of word indexes. # # Note that Whole Word Masking does *not* change the training code # at all -- we still predict each WordPiece independently, softmaxed # over the entire vocabulary. if self.do_whole_word_mask and len(cand_indexes) >= 1 and token.startswith("##"): cand_indexes[-1].append(i) else: cand_indexes.append([i]) random.shuffle(cand_indexes) output_tokens = [t[2:] if len(self.whole_rec.findall(t)) > 0 else t for t in tokens] # 去掉"##" num_to_predict = min(self.max_predictions_per_seq, max(1, int(round(len(tokens) * self.masked_lm_prob)))) masked_lms = [] covered_indexes = set() for index_set in cand_indexes: if len(masked_lms) >= num_to_predict: break # If adding a whole-word mask would exceed the maximum number of # predictions, then just skip this candidate. if len(masked_lms) + len(index_set) > num_to_predict: continue is_any_index_covered = False for index in index_set: if index in covered_indexes: is_any_index_covered = True break if is_any_index_covered: continue for index in index_set: covered_indexes.add(index) masked_token = None # 80% of the time, replace with [MASK] if random.random() < 0.8: masked_token = "[MASK]" else: # 10% of the time, keep original if random.random() < 0.5: masked_token = ( tokens[index][2:] if len(self.whole_rec.findall(tokens[index])) > 0 else tokens[index] ) # 去掉"##" # 10% of the time, replace with random word else: masked_token = self.vocab_words[random.randint(0, len(self.vocab_words) - 1)] output_tokens[index] = masked_token masked_lms.append( MaskedLMInstance( index=index, label=tokens[index][2:] if len(self.whole_rec.findall(tokens[index])) > 0 else tokens[index], ) ) assert len(masked_lms) <= num_to_predict masked_lms = sorted(masked_lms, key=lambda x: x.index) masked_lm_output = [-1] * len(output_tokens) for p in masked_lms: masked_lm_output[p.index] = self.tokenizer.vocab[p.label] return (output_tokens, masked_lm_output)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/community/roberta/preprocessing/sentence_split.py
examples/community/roberta/preprocessing/sentence_split.py
import argparse import functools import json import multiprocessing import os import re import time from typing import List from tqdm import tqdm def split_sentence(document: str, flag: str = "all", limit: int = 510) -> List[str]: sent_list = [] try: if flag == "zh": document = re.sub("(?P<quotation_mark>([。?!…](?![”’\"'])))", r"\g<quotation_mark>\n", document) document = re.sub("(?P<quotation_mark>([。?!]|…{1,2})[”’\"'])", r"\g<quotation_mark>\n", document) elif flag == "en": document = re.sub("(?P<quotation_mark>([.?!](?![”’\"'])))", r"\g<quotation_mark>\n", document) document = re.sub( "(?P<quotation_mark>([?!.][\"']))", r"\g<quotation_mark>\n", document ) # Special quotation marks else: document = re.sub("(?P<quotation_mark>([。?!….?!](?![”’\"'])))", r"\g<quotation_mark>\n", document) document = re.sub( "(?P<quotation_mark>(([。?!.!?]|…{1,2})[”’\"']))", r"\g<quotation_mark>\n", document ) # Special quotation marks sent_list_ori = document.splitlines() for sent in sent_list_ori: sent = sent.strip() if not sent: continue elif len(sent) <= 2: continue else: while len(sent) > limit: temp = sent[0:limit] sent_list.append(temp) sent = sent[limit:] sent_list.append(sent) except: sent_list.clear() sent_list.append(document) return sent_list def get_sent(output_path, input_path, fin_list=[], host=-1, seq_len=512) -> None: workers = 32 if input_path[-1] == "/": input_path = input_path[:-1] cur_path = os.path.join(output_path, str(host) + ".txt") new_split_sentence = functools.partial(split_sentence, limit=seq_len - 2) with open(cur_path, "w", encoding="utf-8") as f: for fi, fin_path in enumerate(fin_list): if not os.path.exists(os.path.join(input_path, fin_path[0])): continue if ".json" not in fin_path[0]: continue print("Processing ", fin_path[0], " ", fi) with open(os.path.join(input_path, fin_path[0]), "r") as fin: f_data = [l["content"] for l in json.load(fin)] pool = multiprocessing.Pool(workers) all_sent = pool.imap_unordered(new_split_sentence, f_data, 32) pool.close() print("finished..") cnt = 0 for d in tqdm(all_sent): for i in d: f.write(i.strip() + "\n") f.write("]]" + "\n") cnt += 1 # if cnt >= 2: # exit() def getFileSize(filepath, shard): all_data = [] for i in os.listdir(filepath): all_data.append(os.path.join(filepath, i)) all_size = sum([os.path.getsize(os.path.join(filepath, f)) for f in all_data]) ans = [[f.split("/")[-1], os.path.getsize(os.path.join(filepath, f))] for f in all_data] ans = sorted(ans, key=lambda x: x[1], reverse=True) per_size = all_size / shard real_shard = [] temp = [] accu_size = 0 for i in ans: accu_size += i[1] temp.append(i) if accu_size > per_size: real_shard.append(temp) accu_size = 0 temp = [] if len(temp) > 0: real_shard.append(temp) return real_shard def get_start_end(real_shard, base=0, server_num=10, server_name="GPU"): import socket host = int(socket.gethostname().split(server_name)[-1]) fin_list = real_shard[server_num * base + host - 1] print(fin_list) print(f"I am server {host}, process {server_num * base + host - 1}, len {len(fin_list)}") return fin_list, host if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--server_num", type=int, default=10, help="number of servers") parser.add_argument("--seq_len", type=int, default=512, help="sequence length") parser.add_argument("--shard", type=int, default=100, help="number of shards, e.g., 10, 50, or 100") parser.add_argument("--input_path", type=str, required=True, help="input path of original corpus") parser.add_argument("--output_path", type=str, required=True, help="output path of shard which has split sentence") args = parser.parse_args() server_num = args.server_num seq_len = args.seq_len shard = args.shard input_path = args.input_path output_path = args.output_path real_shard = getFileSize(input_path, shard) start = time.time() for index, shard in enumerate(real_shard): get_sent(output_path, input_path, fin_list=shard, host=index, seq_len=seq_len) print(f"cost {str(time.time() - start)}") # if you have multiple server, you can use code below or modify code to openmpi # for i in range(len(real_shard) // server_num + 1): # fin_list, host = get_start_end(real_shard, i) # start = time.time() # get_sent(output_path, # input_path, # fin_list=fin_list, host= 10 * i + host - 1) # print(f'cost {str(time.time() - start)}')
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/download_cifar10.py
examples/tutorial/download_cifar10.py
import os from torchvision.datasets import CIFAR10 def main(): dir_path = os.path.dirname(os.path.realpath(__file__)) data_root = os.path.join(dir_path, "data") dataset = CIFAR10(root=data_root, download=True) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/large_batch_optimizer/train.py
examples/tutorial/large_batch_optimizer/train.py
import torch import torch.nn as nn from torchvision.models import resnet18 from tqdm import tqdm import colossalai from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.nn.optimizer import Lamb, Lars class DummyDataloader: def __init__(self, length, batch_size): self.length = length self.batch_size = batch_size def generate(self): data = torch.rand(self.batch_size, 3, 224, 224) label = torch.randint(low=0, high=10, size=(self.batch_size,)) return data, label def __iter__(self): self.step = 0 return self def __next__(self): if self.step < self.length: self.step += 1 return self.generate() else: raise StopIteration def __len__(self): return self.length def main(): # initialize distributed setting parser = colossalai.legacy.get_default_parser() parser.add_argument( "--optimizer", choices=["lars", "lamb"], help="Choose your large-batch optimizer", required=True ) args = parser.parse_args() # launch from torch colossalai.legacy.launch_from_torch(config=args.config) # get logger logger = get_dist_logger() logger.info("initialized distributed environment", ranks=[0]) # create synthetic dataloaders train_dataloader = DummyDataloader(length=10, batch_size=gpc.config.BATCH_SIZE) test_dataloader = DummyDataloader(length=5, batch_size=gpc.config.BATCH_SIZE) # build model model = resnet18(num_classes=gpc.config.NUM_CLASSES) # create loss function criterion = nn.CrossEntropyLoss() # create optimizer if args.optimizer == "lars": optim_cls = Lars elif args.optimizer == "lamb": optim_cls = Lamb optimizer = optim_cls(model.parameters(), lr=gpc.config.LEARNING_RATE, weight_decay=gpc.config.WEIGHT_DECAY) # create lr scheduler lr_scheduler = CosineAnnealingWarmupLR( optimizer=optimizer, total_steps=gpc.config.NUM_EPOCHS, warmup_steps=gpc.config.WARMUP_EPOCHS ) # initialize engine, train_dataloader, test_dataloader, _ = colossalai.legacy.initialize( model=model, optimizer=optimizer, criterion=criterion, train_dataloader=train_dataloader, test_dataloader=test_dataloader, ) logger.info("Engine is built", ranks=[0]) for epoch in range(gpc.config.NUM_EPOCHS): # training engine.train() data_iter = iter(train_dataloader) if gpc.get_global_rank() == 0: description = "Epoch {} / {}".format(epoch, gpc.config.NUM_EPOCHS) progress = tqdm(range(len(train_dataloader)), desc=description) else: progress = range(len(train_dataloader)) for _ in progress: engine.zero_grad() engine.execute_schedule(data_iter, return_output_label=False) engine.step() lr_scheduler.step() if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/large_batch_optimizer/config.py
examples/tutorial/large_batch_optimizer/config.py
from colossalai.legacy.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 512 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config NUM_CLASSES = 10 fp16 = dict(mode=AMP_TYPE.NAIVE) clip_grad_norm = 1.0
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/hybrid_parallel/train.py
examples/tutorial/hybrid_parallel/train.py
import os import torch from titans.model.vit.vit import _create_vit_model from tqdm import tqdm import colossalai from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.nn import CrossEntropyLoss from colossalai.legacy.pipeline.pipelinable import PipelinableContext from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingWarmupLR from colossalai.utils import is_using_pp class DummyDataloader: def __init__(self, length, batch_size): self.length = length self.batch_size = batch_size def generate(self): data = torch.rand(self.batch_size, 3, 224, 224) label = torch.randint(low=0, high=10, size=(self.batch_size,)) return data, label def __iter__(self): self.step = 0 return self def __next__(self): if self.step < self.length: self.step += 1 return self.generate() else: raise StopIteration def __len__(self): return self.length def main(): # launch from torch parser = colossalai.legacy.get_default_parser() args = parser.parse_args() colossalai.legacy.launch_from_torch(config=args.config) # get logger logger = get_dist_logger() logger.info("initialized distributed environment", ranks=[0]) if hasattr(gpc.config, "LOG_PATH"): if gpc.get_global_rank() == 0: log_path = gpc.config.LOG_PATH if not os.path.exists(log_path): os.mkdir(log_path) logger.log_to_file(log_path) use_pipeline = is_using_pp() # create model model_kwargs = dict( img_size=gpc.config.IMG_SIZE, patch_size=gpc.config.PATCH_SIZE, hidden_size=gpc.config.HIDDEN_SIZE, depth=gpc.config.DEPTH, num_heads=gpc.config.NUM_HEADS, mlp_ratio=gpc.config.MLP_RATIO, num_classes=10, init_method="jax", checkpoint=gpc.config.CHECKPOINT, ) if use_pipeline: pipelinable = PipelinableContext() with pipelinable: model = _create_vit_model(**model_kwargs) pipelinable.to_layer_list() pipelinable.policy = "uniform" model = pipelinable.partition(1, gpc.pipeline_parallel_size, gpc.get_local_rank(ParallelMode.PIPELINE)) else: model = _create_vit_model(**model_kwargs) # count number of parameters total_numel = 0 for p in model.parameters(): total_numel += p.numel() if not gpc.is_initialized(ParallelMode.PIPELINE): pipeline_stage = 0 else: pipeline_stage = gpc.get_local_rank(ParallelMode.PIPELINE) logger.info(f"number of parameters: {total_numel} on pipeline stage {pipeline_stage}") # use synthetic dataset # we train for 10 steps and eval for 5 steps per epoch train_dataloader = DummyDataloader(length=10, batch_size=gpc.config.BATCH_SIZE) test_dataloader = DummyDataloader(length=5, batch_size=gpc.config.BATCH_SIZE) # create loss function criterion = CrossEntropyLoss(label_smoothing=0.1) # create optimizer optimizer = torch.optim.AdamW(model.parameters(), lr=gpc.config.LEARNING_RATE, weight_decay=gpc.config.WEIGHT_DECAY) # create lr scheduler lr_scheduler = CosineAnnealingWarmupLR( optimizer=optimizer, total_steps=gpc.config.NUM_EPOCHS, warmup_steps=gpc.config.WARMUP_EPOCHS ) # initialize engine, train_dataloader, test_dataloader, _ = colossalai.initialize( model=model, optimizer=optimizer, criterion=criterion, train_dataloader=train_dataloader, test_dataloader=test_dataloader, ) logger.info("Engine is built", ranks=[0]) for epoch in range(gpc.config.NUM_EPOCHS): # training engine.train() data_iter = iter(train_dataloader) if gpc.get_global_rank() == 0: description = "Epoch {} / {}".format(epoch, gpc.config.NUM_EPOCHS) progress = tqdm(range(len(train_dataloader)), desc=description) else: progress = range(len(train_dataloader)) for _ in progress: engine.zero_grad() engine.execute_schedule(data_iter, return_output_label=False) engine.step() lr_scheduler.step() gpc.destroy() if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/hybrid_parallel/config.py
examples/tutorial/hybrid_parallel/config.py
from colossalai.legacy.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 4 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 2 WARMUP_EPOCHS = 1 # model config IMG_SIZE = 224 PATCH_SIZE = 16 HIDDEN_SIZE = 128 DEPTH = 4 NUM_HEADS = 4 MLP_RATIO = 2 NUM_CLASSES = 10 CHECKPOINT = False SEQ_LENGTH = (IMG_SIZE // PATCH_SIZE) ** 2 + 1 # add 1 for cls token # parallel setting TENSOR_PARALLEL_SIZE = 2 TENSOR_PARALLEL_MODE = "1d" parallel = dict( pipeline=2, tensor=dict(mode=TENSOR_PARALLEL_MODE, size=TENSOR_PARALLEL_SIZE), ) fp16 = dict(mode=AMP_TYPE.NAIVE) clip_grad_norm = 1.0 # pipeline config NUM_MICRO_BATCHES = parallel["pipeline"]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/colossalai_zero.py
examples/tutorial/opt/opt/colossalai_zero.py
try: from colossalai.zero.shard_utils import TensorShardStrategy except ImportError: # colossalai > 0.2.8 from colossalai.legacy.zero import TensorShardStrategy zero = dict( model_config=dict(shard_strategy=TensorShardStrategy(), tensor_placement_policy="auto", reuse_fp16_shard=True), optimizer_config=dict(gpu_margin_mem_ratio=0.8, initial_scale=16384), )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/context.py
examples/tutorial/opt/opt/context.py
import torch.distributed as dist from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc class barrier_context: """ This context manager is used to allow one process to execute while blocking all other processes in the same process group. This is often useful when downloading is required as we only want to download in one process to prevent file corruption. Args: executor_rank (int): the process rank to execute without blocking, all other processes will be blocked parallel_mode (ParallelMode): the parallel mode corresponding to a process group Usage: with barrier_context(): dataset = CIFAR10(root='./data', download=True) """ def __init__(self, executor_rank: int = 0, parallel_mode: ParallelMode = ParallelMode.GLOBAL): # the class name is lowercase by convention current_rank = gpc.get_local_rank(parallel_mode=parallel_mode) self.should_block = current_rank != executor_rank self.group = gpc.get_group(parallel_mode=parallel_mode) def __enter__(self): if self.should_block: dist.barrier(group=self.group) def __exit__(self, exc_type, exc_value, exc_traceback): if not self.should_block: dist.barrier(group=self.group)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/opt/run_clm.py
examples/tutorial/opt/opt/run_clm.py
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """ Fine-tuning the library models for causal language modeling (GPT, GPT-2, CTRL, ...) on a text file or a dataset without using HuggingFace Trainer. Here is the full list of checkpoints on the hub that can be fine-tuned by this script: https://huggingface.co/models?filter=text-generation """ # You can also adapt this script on your own causal language modeling task. Pointers for this are left as comments. import math import os import time from itertools import chain import datasets import torch import torch.distributed as dist import transformers.utils.logging as logging from accelerate.utils import set_seed from context import barrier_context from datasets import load_dataset from packaging import version from torch.utils.data import DataLoader from tqdm.auto import tqdm from transformers import ( CONFIG_MAPPING, MODEL_MAPPING, AutoConfig, AutoTokenizer, GPT2Tokenizer, OPTForCausalLM, SchedulerType, default_data_collator, get_scheduler, ) from transformers.utils.versions import require_version import colossalai from colossalai.accelerator import get_accelerator from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.tensor import ProcessGroup from colossalai.legacy.utils import get_dataloader from colossalai.logging import disable_existing_loggers, get_dist_logger from colossalai.nn.optimizer import HybridAdam from colossalai.zero import GeminiOptimizer require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/language-modeling/requirements.txt") MODEL_CONFIG_CLASSES = list(MODEL_MAPPING.keys()) MODEL_TYPES = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) def get_time_stamp(): torch.cuda.synchronize() return time.time() def parse_args(): parser = colossalai.legacy.get_default_parser() parser.add_argument("-s", "--synthetic", action="store_true") parser.add_argument( "--dataset_name", type=str, default=None, help="The name of the dataset to use (via the datasets library).", ) parser.add_argument( "--dataset_config_name", type=str, default=None, help="The configuration name of the dataset to use (via the datasets library).", ) parser.add_argument( "--train_file", type=str, default=None, help="A csv or a json file containing the training data." ) parser.add_argument( "--validation_file", type=str, default=None, help="A csv or a json file containing the validation data." ) parser.add_argument( "--validation_split_percentage", default=5, help="The percentage of the train set used as validation set in case there's no validation split", ) parser.add_argument( "--model_name_or_path", type=str, help="Path to pretrained model or model identifier from huggingface.co/models.", required=True, ) parser.add_argument( "--config_name", type=str, default=None, help="Pretrained config name or path if not the same as model_name", ) parser.add_argument( "--tokenizer_name", type=str, default=None, help="Pretrained tokenizer name or path if not the same as model_name", ) parser.add_argument( "--use_slow_tokenizer", action="store_true", help="If passed, will use a slow tokenizer (not backed by the 🤗 Tokenizers library).", ) parser.add_argument( "--per_device_train_batch_size", type=int, default=8, help="Batch size (per device) for the training dataloader.", ) parser.add_argument( "--per_device_eval_batch_size", type=int, default=8, help="Batch size (per device) for the evaluation dataloader.", ) parser.add_argument( "--learning_rate", type=float, default=5e-5, help="Initial learning rate (after the potential warmup period) to use.", ) parser.add_argument("--weight_decay", type=float, default=0.0, help="Weight decay to use.") parser.add_argument("--num_train_epochs", type=int, default=3, help="Total number of training epochs to perform.") parser.add_argument( "--max_train_steps", type=int, default=None, help="Total number of training steps to perform. If provided, overrides num_train_epochs.", ) parser.add_argument( "--gradient_accumulation_steps", type=int, default=1, help="Number of updates steps to accumulate before performing a backward/update pass.", ) parser.add_argument( "--lr_scheduler_type", type=SchedulerType, default="linear", help="The scheduler type to use.", choices=["linear", "cosine", "cosine_with_restarts", "polynomial", "constant", "constant_with_warmup"], ) parser.add_argument( "--num_warmup_steps", type=int, default=0, help="Number of steps for the warmup in the lr scheduler." ) parser.add_argument("--output_dir", type=str, default=None, help="Where to store the final model.") parser.add_argument("--seed", type=int, default=None, help="A seed for reproducible training.") parser.add_argument( "--model_type", type=str, default=None, help="Model type to use if training from scratch.", choices=MODEL_TYPES, ) parser.add_argument( "--block_size", type=int, default=None, help=( "Optional input sequence length after tokenization. The training dataset will be truncated in block of" " this size for training. Default to the model max input length for single sentence inputs (take into" " account special tokens)." ), ) parser.add_argument( "--preprocessing_num_workers", type=int, default=None, help="The number of processes to use for the preprocessing.", ) parser.add_argument( "--overwrite_cache", type=bool, default=False, help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--no_keep_linebreaks", action="store_true", help="Do not keep line breaks when using TXT files." ) parser.add_argument("--push_to_hub", action="store_true", help="Whether or not to push the model to the Hub.") parser.add_argument( "--hub_model_id", type=str, help="The name of the repository to keep in sync with the local `output_dir`." ) parser.add_argument("--hub_token", type=str, help="The token to use to push to the Model Hub.") parser.add_argument( "--checkpointing_steps", type=str, default=None, help="Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch.", ) parser.add_argument( "--resume_from_checkpoint", type=str, default=None, help="If the training should continue from a checkpoint folder.", ) parser.add_argument( "--with_tracking", action="store_true", help="Whether to enable experiment trackers for logging.", ) parser.add_argument( "--report_to", type=str, default="all", help=( 'The integration to report the results and logs to. Supported platforms are `"tensorboard"`,' ' `"wandb"` and `"comet_ml"`. Use `"all"` (default) to report to all integrations.' "Only applicable when `--with_tracking` is passed." ), ) parser.add_argument("--mem_cap", type=int, default=0, help="use mem cap") parser.add_argument("--init_in_cpu", action="store_true", default=False, help="init training model in cpu") args = parser.parse_args() # Sanity checks if not args.synthetic: if args.dataset_name is None and args.train_file is None and args.validation_file is None: raise ValueError("Need either a dataset name or a training/validation file.") else: if args.train_file is not None: extension = args.train_file.split(".")[-1] assert extension in ["csv", "json", "txt"], "`train_file` should be a csv, json or txt file." if args.validation_file is not None: extension = args.validation_file.split(".")[-1] assert extension in ["csv", "json", "txt"], "`validation_file` should be a csv, json or txt file." if args.push_to_hub: assert args.output_dir is not None, "Need an `output_dir` to create a repo when `--push_to_hub` is passed." return args def colo_memory_cap(size_in_GB): from colossalai.utils import colo_device_memory_capacity, colo_set_process_memory_fraction cuda_capacity = colo_device_memory_capacity(get_accelerator().get_current_device()) if size_in_GB * (1024**3) < cuda_capacity: colo_set_process_memory_fraction(size_in_GB * (1024**3) / cuda_capacity) print("Using {} GB of GPU memory".format(size_in_GB)) class DummyDataloader: def __init__(self, length, batch_size, seq_len, vocab_size): self.length = length self.batch_size = batch_size self.seq_len = seq_len self.vocab_size = vocab_size def generate(self): input_ids = torch.randint( 0, self.vocab_size, (self.batch_size, self.seq_len), device=get_accelerator().get_current_device() ) attention_mask = torch.ones_like(input_ids) return {"input_ids": input_ids, "attention_mask": attention_mask, "labels": input_ids} def __iter__(self): self.step = 0 return self def __next__(self): if self.step < self.length: self.step += 1 return self.generate() else: raise StopIteration def __len__(self): return self.length def main(): args = parse_args() disable_existing_loggers() colossalai.legacy.launch_from_torch() logger = get_dist_logger() is_main_process = dist.get_rank() == 0 if is_main_process: datasets.utils.logging.set_verbosity_warning() logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() logging.set_verbosity_error() if args.mem_cap > 0: colo_memory_cap(args.mem_cap) # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) logger.info(f"Rank {dist.get_rank()}: random seed is set to {args.seed}") # Handle the repository creation with barrier_context(): if args.output_dir is not None: os.makedirs(args.output_dir, exist_ok=True) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. logger.info("Start preparing dataset", ranks=[0]) if not args.synthetic: if args.dataset_name is not None: # Downloading and loading a dataset from the hub. raw_datasets = load_dataset(args.dataset_name, args.dataset_config_name) if "validation" not in raw_datasets.keys(): raw_datasets["validation"] = load_dataset( args.dataset_name, args.dataset_config_name, split=f"train[:{args.validation_split_percentage}%]", ) raw_datasets["train"] = load_dataset( args.dataset_name, args.dataset_config_name, split=f"train[{args.validation_split_percentage}%:]", ) else: data_files = {} dataset_args = {} if args.train_file is not None: data_files["train"] = args.train_file if args.validation_file is not None: data_files["validation"] = args.validation_file extension = args.train_file.split(".")[-1] if extension == "txt": extension = "text" dataset_args["keep_linebreaks"] = not args.no_keep_linebreaks raw_datasets = load_dataset(extension, data_files=data_files, **dataset_args) # If no validation data is there, validation_split_percentage will be used to divide the dataset. if "validation" not in raw_datasets.keys(): raw_datasets["validation"] = load_dataset( extension, data_files=data_files, split=f"train[:{args.validation_split_percentage}%]", **dataset_args, ) raw_datasets["train"] = load_dataset( extension, data_files=data_files, split=f"train[{args.validation_split_percentage}%:]", **dataset_args, ) logger.info("Dataset is prepared", ranks=[0]) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if args.config_name: config = AutoConfig.from_pretrained(args.config_name) elif args.model_name_or_path: config = AutoConfig.from_pretrained(args.model_name_or_path) else: config = CONFIG_MAPPING[args.model_type]() logger.warning("You are instantiating a new config instance from scratch.") logger.info("Model config has been created", ranks=[0]) if args.model_name_or_path == "facebook/opt-13b": tokenizer = GPT2Tokenizer.from_pretrained(args.model_name_or_path) else: print(f"load model from {args.model_name_or_path}") tokenizer = AutoTokenizer.from_pretrained(args.model_name_or_path, use_fast=not args.use_slow_tokenizer) logger.info(f"{tokenizer.__class__.__name__} has been created", ranks=[0]) if args.init_in_cpu: init_dev = torch.device("cpu") else: init_dev = get_accelerator().get_current_device() cai_version = colossalai.__version__ logger.info(f"using Colossal-AI version {cai_version}") # build model if version.parse(cai_version) >= version.parse("0.3.1"): from contextlib import nullcontext from colossalai.lazy import LazyInitContext ctx = ( LazyInitContext(default_device=init_dev) if args.model_name_or_path is None or args.model_name_or_path == "facebook/opt-13b" else nullcontext() ) else: from colossalai.zero import ColoInitContext ctx = ColoInitContext(device=init_dev) if args.model_name_or_path is None or args.model_name_or_path == "facebook/opt-13b": # currently, there has a bug in pretrained opt-13b # we can not import it until huggingface fix it logger.info("Train a new model from scratch", ranks=[0]) with ctx: model = OPTForCausalLM(config) else: logger.info("Finetune a pre-trained model", ranks=[0]) with ctx: model = OPTForCausalLM.from_pretrained( args.model_name_or_path, from_tf=bool(".ckpt" in args.model_name_or_path), config=config, local_files_only=False, ) # enable graident checkpointing model.gradient_checkpointing_enable() PLACEMENT_POLICY = "auto" if version.parse(cai_version) >= version.parse("0.3.1"): from colossalai.zero import GeminiDDP model = GeminiDDP(model, offload_optim_frac=1.0, pin_memory=True) elif version.parse(cai_version) > version.parse("0.1.10"): try: from colossalai.nn.parallel import GeminiDDP except ImportError: # this works for unreleased main branch, and this may be released on 0.2.9 from colossalai.zero import GeminiDDP model = GeminiDDP( model, device=get_accelerator().get_current_device(), placement_policy=PLACEMENT_POLICY, pin_memory=True ) elif version.parse(cai_version) <= version.parse("0.1.10") and version.parse(cai_version) >= version.parse("0.1.9"): from colossalai.gemini import ChunkManager, GeminiManager pg = ProcessGroup() chunk_size = ChunkManager.search_chunk_size(model, 64 * 1024**2, 32) chunk_manager = ChunkManager( chunk_size, pg, enable_distributed_storage=True, init_device=GeminiManager.get_default_device(PLACEMENT_POLICY), ) gemini_manager = GeminiManager(PLACEMENT_POLICY, chunk_manager) model = ZeroDDP(model, gemini_manager) logger.info(f"{model.__class__.__name__} has been created", ranks=[0]) if not args.synthetic: # Preprocessing the datasets. # First we tokenize all the texts. column_names = raw_datasets["train"].column_names text_column_name = "text" if "text" in column_names else column_names[0] def tokenize_function(examples): return tokenizer(examples[text_column_name]) with barrier_context(executor_rank=0, parallel_mode=ParallelMode.DATA): tokenized_datasets = raw_datasets.map( tokenize_function, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc="Running tokenizer on dataset", ) if args.block_size is None: block_size = tokenizer.model_max_length if block_size > 1024: logger.warning( f"The tokenizer picked seems to have a very large `model_max_length` ({tokenizer.model_max_length}). " "Picking 1024 instead. You can change that default value by passing --block_size xxx." ) block_size = 1024 else: if args.block_size > tokenizer.model_max_length: logger.warning( f"The block_size passed ({args.block_size}) is larger than the maximum length for the model" f"({tokenizer.model_max_length}). Using block_size={tokenizer.model_max_length}." ) block_size = min(args.block_size, tokenizer.model_max_length) # Main data processing function that will concatenate all texts from our dataset and generate chunks of block_size. def group_texts(examples): # Concatenate all texts. concatenated_examples = {k: list(chain(*examples[k])) for k in examples.keys()} total_length = len(concatenated_examples[list(examples.keys())[0]]) # We drop the small remainder, we could add padding if the model supported it instead of this drop, you can # customize this part to your needs. if total_length >= block_size: total_length = (total_length // block_size) * block_size # Split by chunks of max_len. result = { k: [t[i : i + block_size] for i in range(0, total_length, block_size)] for k, t in concatenated_examples.items() } result["labels"] = result["input_ids"].copy() return result if not args.synthetic: # Note that with `batched=True`, this map processes 1,000 texts together, so group_texts throws away a remainder # for each of those groups of 1,000 texts. You can adjust that batch_size here but a higher value might be slower # to preprocess. # # To speed up this part, we use multiprocessing. See the documentation of the map method for more information: # https://huggingface.co/docs/datasets/package_reference/main_classes.html#datasets.Dataset.map with barrier_context(executor_rank=0, parallel_mode=ParallelMode.DATA): lm_datasets = tokenized_datasets.map( group_texts, batched=True, num_proc=args.preprocessing_num_workers, load_from_cache_file=not args.overwrite_cache, desc=f"Grouping texts in chunks of {block_size}", ) train_dataset = lm_datasets["train"] eval_dataset = lm_datasets["validation"] # Log a few random samples from the training set: # for index in random.sample(range(len(train_dataset)), 3): # logger.info(f"Sample {index} of the training set: {train_dataset[index]}.") # DataLoaders creation: train_dataloader = get_dataloader( train_dataset, shuffle=True, add_sampler=True, collate_fn=default_data_collator, batch_size=args.per_device_train_batch_size, ) eval_dataloader = DataLoader( eval_dataset, collate_fn=default_data_collator, batch_size=args.per_device_eval_batch_size ) else: train_dataloader = DummyDataloader( 30, args.per_device_train_batch_size, config.max_position_embeddings, config.vocab_size ) eval_dataloader = DummyDataloader( 10, args.per_device_train_batch_size, config.max_position_embeddings, config.vocab_size ) logger.info("Dataloaders have been created", ranks=[0]) # Optimizer # Split weights in two groups, one with weight decay and the other not. no_decay = ["bias", "LayerNorm.weight"] optimizer_grouped_parameters = [ { "params": [p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay)], "weight_decay": args.weight_decay, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], "weight_decay": 0.0, }, ] optimizer = HybridAdam(optimizer_grouped_parameters, lr=args.learning_rate) # Scheduler and math around the number of training steps. overrode_max_train_steps = False num_update_steps_per_epoch = math.ceil(len(train_dataloader) / args.gradient_accumulation_steps) if args.max_train_steps is None: args.max_train_steps = args.num_train_epochs * num_update_steps_per_epoch overrode_max_train_steps = True lr_scheduler = get_scheduler( name=args.lr_scheduler_type, optimizer=optimizer, num_warmup_steps=args.num_warmup_steps, num_training_steps=args.max_train_steps, ) optimizer = GeminiOptimizer(optimizer, model, initial_scale=2**14) # We need to recalculate our total training steps as the size of the training dataloader may have changed. num_update_steps_per_epoch = math.ceil(len(train_dataloader) / args.gradient_accumulation_steps) if overrode_max_train_steps: args.max_train_steps = args.num_train_epochs * num_update_steps_per_epoch # Afterwards we recalculate our number of training epochs args.num_train_epochs = math.ceil(args.max_train_steps / num_update_steps_per_epoch) # Train! total_batch_size = args.per_device_train_batch_size * gpc.get_world_size(ParallelMode.DATA) num_train_samples = len(train_dataset) if not args.synthetic else 30 * total_batch_size num_eval_samples = len(eval_dataset) if not args.synthetic else 10 * total_batch_size logger.info("***** Running training *****", ranks=[0]) logger.info(f" Num examples = {num_train_samples}", ranks=[0]) logger.info(f" Num Epochs = {args.num_train_epochs}", ranks=[0]) logger.info(f" Instantaneous batch size per device = {args.per_device_train_batch_size}", ranks=[0]) logger.info(f" Total train batch size (w. parallel, distributed & accumulation) = {total_batch_size}", ranks=[0]) logger.info(f" Gradient Accumulation steps = {args.gradient_accumulation_steps}", ranks=[0]) logger.info(f" Total optimization steps = {args.max_train_steps}", ranks=[0]) # Only show the progress bar once on each machine. progress_bar = tqdm(range(args.max_train_steps), disable=not is_main_process) completed_steps = 0 starting_epoch = 0 global_step = 0 for epoch in range(starting_epoch, args.num_train_epochs): if completed_steps >= args.max_train_steps: break model.train() for step, batch in enumerate(train_dataloader): batch = {k: v.cuda() for k, v in batch.items()} outputs = model(use_cache=False, **batch) loss = outputs["loss"] optimizer.backward(loss) if step % args.gradient_accumulation_steps == 0 or step == len(train_dataloader) - 1: optimizer.step() lr_scheduler.step() optimizer.zero_grad() progress_bar.update(1) completed_steps += 1 global_step += 1 logger.info("Global step {} finished".format(global_step + 1), ranks=[0]) if completed_steps >= args.max_train_steps: break model.eval() losses = [] for step, batch in enumerate(eval_dataloader): with torch.no_grad(): batch = {k: v.cuda() for k, v in batch.items()} outputs = model(**batch) loss = outputs["loss"].unsqueeze(0) losses.append(loss) losses = torch.cat(losses) losses = losses[:num_eval_samples] try: eval_loss = torch.mean(losses) perplexity = math.exp(eval_loss) except OverflowError: perplexity = float("inf") logger.info(f"Epoch {epoch}: perplexity: {perplexity} eval_loss: {eval_loss}", ranks=[0]) if args.output_dir is not None: model_state = model.state_dict() if is_main_process: torch.save(model_state, args.output_dir + "/epoch_{}_model.pth".format(completed_steps)) dist.barrier() # load_state = torch.load(args.output_dir + '/epoch_{}_model.pth'.format(completed_steps)) # model.load_state_dict(load_state, strict=False) logger.info("Training finished", ranks=[0]) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/opt_server.py
examples/tutorial/opt/inference/opt_server.py
import argparse import logging import random from typing import Optional from batch import BatchManagerForGeneration from cache import ListCache, MissCacheError from energonai import QueueFullError, launch_engine from energonai.model import opt_6B, opt_30B, opt_125M, opt_175B from pydantic import BaseModel, Field from sanic import Sanic from sanic.request import Request from sanic.response import json from sanic_ext import openapi, validate from torch import Tensor from transformers import GPT2Tokenizer class GenerationTaskReq(BaseModel): max_tokens: int = Field(gt=0, le=256, example=64) prompt: str = Field( min_length=1, example="Question: Where were the 2004 Olympics held?\nAnswer: Athens, Greece\n\nQuestion: What is the longest river on the earth?\nAnswer:", ) top_k: Optional[int] = Field(default=None, gt=0, example=50) top_p: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.5) temperature: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.7) app = Sanic("opt") @app.post("/generation") @openapi.body(GenerationTaskReq) @validate(json=GenerationTaskReq) async def generate(request: Request, body: GenerationTaskReq): logger.info(f'{request.ip}:{request.port} - "{request.method} {request.path}" - {body}') key = (body.prompt, body.max_tokens) try: if cache is None: raise MissCacheError() outputs = cache.get(key) output = random.choice(outputs) logger.info("Cache hit") except MissCacheError: inputs = tokenizer(body.prompt, truncation=True, max_length=512) inputs["max_tokens"] = body.max_tokens inputs["top_k"] = body.top_k inputs["top_p"] = body.top_p inputs["temperature"] = body.temperature try: uid = id(body) engine.submit(uid, inputs) output = await engine.wait(uid) assert isinstance(output, Tensor) output = tokenizer.decode(output, skip_special_tokens=True) if cache is not None: cache.add(key, output) except QueueFullError as e: return json({"detail": e.args[0]}, status=406) return json({"text": output}) @app.after_server_stop def shutdown(*_): engine.shutdown() def get_model_fn(model_name: str): model_map = {"opt-125m": opt_125M, "opt-6.7b": opt_6B, "opt-30b": opt_30B, "opt-175b": opt_175B} return model_map[model_name] def print_args(args: argparse.Namespace): print("\n==> Args:") for k, v in args.__dict__.items(): print(f"{k} = {v}") FIXED_CACHE_KEYS = [ ( "Question: What is the name of the largest continent on earth?\nAnswer: Asia\n\nQuestion: What is at the center of the solar system?\nAnswer:", 64, ), ( "A chat between a salesman and a student.\n\nSalesman: Hi boy, are you looking for a new phone?\nStudent: Yes, my phone is not functioning well.\nSalesman: What is your budget? \nStudent: I have received my scholarship so I am fine with any phone.\nSalesman: Great, then perhaps this latest flagship phone is just right for you.", 64, ), ( "English: I am happy today.\nChinese: 我今天很开心。\n\nEnglish: I am going to play basketball.\nChinese: 我一会去打篮球。\n\nEnglish: Let's celebrate our anniversary.\nChinese:", 64, ), ] if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("model", choices=["opt-125m", "opt-6.7b", "opt-30b", "opt-175b"]) parser.add_argument("--tp", type=int, default=1) parser.add_argument("--master_host", default="localhost") parser.add_argument("--master_port", type=int, default=19990) parser.add_argument("--rpc_port", type=int, default=19980) parser.add_argument("--max_batch_size", type=int, default=8) parser.add_argument("--pipe_size", type=int, default=1) parser.add_argument("--queue_size", type=int, default=0) parser.add_argument("--http_host", default="0.0.0.0") parser.add_argument("--http_port", type=int, default=7070) parser.add_argument("--checkpoint", default=None) parser.add_argument("--cache_size", type=int, default=0) parser.add_argument("--cache_list_size", type=int, default=1) args = parser.parse_args() print_args(args) model_kwargs = {} if args.checkpoint is not None: model_kwargs["checkpoint"] = args.checkpoint logger = logging.getLogger(__name__) tokenizer = GPT2Tokenizer.from_pretrained("facebook/opt-30b") if args.cache_size > 0: cache = ListCache(args.cache_size, args.cache_list_size, fixed_keys=FIXED_CACHE_KEYS) else: cache = None engine = launch_engine( args.tp, 1, args.master_host, args.master_port, args.rpc_port, get_model_fn(args.model), batch_manager=BatchManagerForGeneration( max_batch_size=args.max_batch_size, pad_token_id=tokenizer.pad_token_id ), pipe_size=args.pipe_size, queue_size=args.queue_size, **model_kwargs, ) app.run(args.http_host, args.http_port)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/opt_fastapi.py
examples/tutorial/opt/inference/opt_fastapi.py
import argparse import logging import random from typing import Optional import uvicorn from batch import BatchManagerForGeneration from cache import ListCache, MissCacheError from energonai import QueueFullError, launch_engine from energonai.model import opt_6B, opt_30B, opt_125M, opt_175B from fastapi import FastAPI, HTTPException, Request from pydantic import BaseModel, Field from transformers import GPT2Tokenizer class GenerationTaskReq(BaseModel): max_tokens: int = Field(gt=0, le=256, example=64) prompt: str = Field( min_length=1, example="Question: Where were the 2004 Olympics held?\nAnswer: Athens, Greece\n\nQuestion: What is the longest river on the earth?\nAnswer:", ) top_k: Optional[int] = Field(default=None, gt=0, example=50) top_p: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.5) temperature: Optional[float] = Field(default=None, gt=0.0, lt=1.0, example=0.7) app = FastAPI() @app.post("/generation") async def generate(data: GenerationTaskReq, request: Request): logger.info(f'{request.client.host}:{request.client.port} - "{request.method} {request.url.path}" - {data}') key = (data.prompt, data.max_tokens) try: if cache is None: raise MissCacheError() outputs = cache.get(key) output = random.choice(outputs) logger.info("Cache hit") except MissCacheError: inputs = tokenizer(data.prompt, truncation=True, max_length=512) inputs["max_tokens"] = data.max_tokens inputs["top_k"] = data.top_k inputs["top_p"] = data.top_p inputs["temperature"] = data.temperature try: uid = id(data) engine.submit(uid, inputs) output = await engine.wait(uid) output = tokenizer.decode(output, skip_special_tokens=True) if cache is not None: cache.add(key, output) except QueueFullError as e: raise HTTPException(status_code=406, detail=e.args[0]) return {"text": output} @app.on_event("shutdown") async def shutdown(*_): engine.shutdown() server.should_exit = True server.force_exit = True await server.shutdown() def get_model_fn(model_name: str): model_map = {"opt-125m": opt_125M, "opt-6.7b": opt_6B, "opt-30b": opt_30B, "opt-175b": opt_175B} return model_map[model_name] def print_args(args: argparse.Namespace): print("\n==> Args:") for k, v in args.__dict__.items(): print(f"{k} = {v}") FIXED_CACHE_KEYS = [ ( "Question: What is the name of the largest continent on earth?\nAnswer: Asia\n\nQuestion: What is at the center of the solar system?\nAnswer:", 64, ), ( "A chat between a salesman and a student.\n\nSalesman: Hi boy, are you looking for a new phone?\nStudent: Yes, my phone is not functioning well.\nSalesman: What is your budget? \nStudent: I have received my scholarship so I am fine with any phone.\nSalesman: Great, then perhaps this latest flagship phone is just right for you.", 64, ), ( "English: I am happy today.\nChinese: 我今天很开心。\n\nEnglish: I am going to play basketball.\nChinese: 我一会去打篮球。\n\nEnglish: Let's celebrate our anniversary.\nChinese:", 64, ), ] if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("model", choices=["opt-125m", "opt-6.7b", "opt-30b", "opt-175b"]) parser.add_argument("--tp", type=int, default=1) parser.add_argument("--master_host", default="localhost") parser.add_argument("--master_port", type=int, default=19990) parser.add_argument("--rpc_port", type=int, default=19980) parser.add_argument("--max_batch_size", type=int, default=8) parser.add_argument("--pipe_size", type=int, default=1) parser.add_argument("--queue_size", type=int, default=0) parser.add_argument("--http_host", default="0.0.0.0") parser.add_argument("--http_port", type=int, default=7070) parser.add_argument("--checkpoint", default=None) parser.add_argument("--cache_size", type=int, default=0) parser.add_argument("--cache_list_size", type=int, default=1) args = parser.parse_args() print_args(args) model_kwargs = {} if args.checkpoint is not None: model_kwargs["checkpoint"] = args.checkpoint logger = logging.getLogger(__name__) tokenizer = GPT2Tokenizer.from_pretrained("facebook/opt-30b") if args.cache_size > 0: cache = ListCache(args.cache_size, args.cache_list_size, fixed_keys=FIXED_CACHE_KEYS) else: cache = None engine = launch_engine( args.tp, 1, args.master_host, args.master_port, args.rpc_port, get_model_fn(args.model), batch_manager=BatchManagerForGeneration( max_batch_size=args.max_batch_size, pad_token_id=tokenizer.pad_token_id ), pipe_size=args.pipe_size, queue_size=args.queue_size, **model_kwargs, ) config = uvicorn.Config(app, host=args.http_host, port=args.http_port) server = uvicorn.Server(config=config) server.run()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/cache.py
examples/tutorial/opt/inference/cache.py
from collections import OrderedDict from contextlib import contextmanager from threading import Lock from typing import Any, Dict, Hashable, List class MissCacheError(Exception): pass class ListCache: def __init__(self, cache_size: int, list_size: int, fixed_keys: List[Hashable] = []) -> None: """Cache a list of values. The fixed keys won't be removed. For other keys, LRU is applied. When the value list is not full, a cache miss occurs. Otherwise, a cache hit occurs. Redundant values will be removed. Args: cache_size (int): Max size for LRU cache. list_size (int): Value list size. fixed_keys (List[Hashable], optional): The keys which won't be removed. Defaults to []. """ self.cache_size = cache_size self.list_size = list_size self.cache: OrderedDict[Hashable, List[Any]] = OrderedDict() self.fixed_cache: Dict[Hashable, List[Any]] = {} for key in fixed_keys: self.fixed_cache[key] = [] self._lock = Lock() def get(self, key: Hashable) -> List[Any]: with self.lock(): if key in self.fixed_cache: l = self.fixed_cache[key] if len(l) >= self.list_size: return l elif key in self.cache: self.cache.move_to_end(key) l = self.cache[key] if len(l) >= self.list_size: return l raise MissCacheError() def add(self, key: Hashable, value: Any) -> None: with self.lock(): if key in self.fixed_cache: l = self.fixed_cache[key] if len(l) < self.list_size and value not in l: l.append(value) elif key in self.cache: self.cache.move_to_end(key) l = self.cache[key] if len(l) < self.list_size and value not in l: l.append(value) else: if len(self.cache) >= self.cache_size: self.cache.popitem(last=False) self.cache[key] = [value] @contextmanager def lock(self): try: self._lock.acquire() yield finally: self._lock.release()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/batch.py
examples/tutorial/opt/inference/batch.py
from typing import Any, Deque, Hashable, List, Tuple import torch from energonai import BatchManager, SubmitEntry, TaskEntry class BatchManagerForGeneration(BatchManager): def __init__(self, max_batch_size: int = 1, pad_token_id: int = 0) -> None: super().__init__() self.max_batch_size = max_batch_size self.pad_token_id = pad_token_id def _left_padding(self, batch_inputs): max_len = max(len(inputs["input_ids"]) for inputs in batch_inputs) outputs = {"input_ids": [], "attention_mask": []} for inputs in batch_inputs: input_ids, attention_mask = inputs["input_ids"], inputs["attention_mask"] padding_len = max_len - len(input_ids) input_ids = [self.pad_token_id] * padding_len + input_ids attention_mask = [0] * padding_len + attention_mask outputs["input_ids"].append(input_ids) outputs["attention_mask"].append(attention_mask) for k in outputs: outputs[k] = torch.tensor(outputs[k]) return outputs, max_len @staticmethod def _make_batch_key(entry: SubmitEntry) -> tuple: data = entry.data return (data["top_k"], data["top_p"], data["temperature"]) def make_batch(self, q: Deque[SubmitEntry]) -> Tuple[TaskEntry, dict]: entry = q.popleft() uids = [entry.uid] batch = [entry.data] while len(batch) < self.max_batch_size: if len(q) == 0: break if self._make_batch_key(entry) != self._make_batch_key(q[0]): break if q[0].data["max_tokens"] > entry.data["max_tokens"]: break e = q.popleft() batch.append(e.data) uids.append(e.uid) inputs, max_len = self._left_padding(batch) trunc_lens = [] for data in batch: trunc_lens.append(max_len + data["max_tokens"]) inputs["top_k"] = entry.data["top_k"] inputs["top_p"] = entry.data["top_p"] inputs["temperature"] = entry.data["temperature"] inputs["max_tokens"] = max_len + entry.data["max_tokens"] return TaskEntry(tuple(uids), inputs), {"trunc_lens": trunc_lens} def split_batch(self, task_entry: TaskEntry, trunc_lens: List[int] = []) -> List[Tuple[Hashable, Any]]: retval = [] for uid, output, trunc_len in zip(task_entry.uids, task_entry.batch, trunc_lens): retval.append((uid, output[:trunc_len])) return retval
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/benchmark/locustfile.py
examples/tutorial/opt/inference/benchmark/locustfile.py
from locust import HttpUser, task class GenerationUser(HttpUser): @task def generate(self): prompt = "Question: What is the longest river on the earth? Answer:" for i in range(4, 9): data = {"max_tokens": 2**i, "prompt": prompt} with self.client.post("/generation", json=data, catch_response=True) as response: if response.status_code in (200, 406): response.success() else: response.failure("Response wrong")
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/script/processing_ckpt_66b.py
examples/tutorial/opt/inference/script/processing_ckpt_66b.py
import os from multiprocessing import Pool import torch # download pytorch model ckpt in https://huggingface.co/facebook/opt-66b/tree/main # you can use whether wget or git lfs path = "/path/to/your/ckpt" new_path = "/path/to/the/processed/ckpt/" assert os.path.isdir(path) files = [] for filename in os.listdir(path): filepath = os.path.join(path, filename) if os.path.isfile(filepath): files.append(filepath) with Pool(14) as pool: ckpts = pool.map(torch.load, files) restored = {} for ckpt in ckpts: for k, v in ckpt.items(): if k[0] == "m": k = k[6:] if k == "lm_head.weight": k = "head.dense.weight" if k == "decoder.final_layer_norm.weight": k = "decoder.layer_norm.weight" if k == "decoder.final_layer_norm.bias": k = "decoder.layer_norm.bias" restored[k] = v restored["decoder.version"] = "0.0" split_num = len(restored.keys()) // 60 count = 0 file_count = 1 tmp = {} for k, v in restored.items(): print(k) tmp[k] = v count = count + 1 if count == split_num: filename = str(file_count) + "-restored.pt" torch.save(tmp, os.path.join(new_path, filename)) file_count = file_count + 1 count = 0 tmp = {} filename = str(file_count) + "-restored.pt" torch.save(tmp, os.path.join(new_path, filename))
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/opt/inference/script/process-opt-175b/convert_ckpt.py
examples/tutorial/opt/inference/script/process-opt-175b/convert_ckpt.py
import argparse import json import os import re from collections import defaultdict import numpy as np import torch def load_json(path: str): with open(path) as f: return json.load(f) def parse_shape_info(flat_dir: str): data = load_json(os.path.join(flat_dir, "shape.json")) flat_info = defaultdict(lambda: defaultdict(list)) for k, shape in data.items(): matched = re.match(r"decoder.layers.\d+", k) if matched is None: flat_key = "flat_param_0" else: flat_key = f"{matched[0]}.flat_param_0" flat_info[flat_key]["names"].append(k) flat_info[flat_key]["shapes"].append(shape) flat_info[flat_key]["numels"].append(int(np.prod(shape))) return flat_info def convert(flat_dir: str, output_dir: str, part: int): flat_path = os.path.join(flat_dir, f"reshard-model_part-{part}-shard0.pt") output_path = os.path.join(output_dir, f"reshard-model_part-{part}.pt") flat_meta = load_json(os.path.join(flat_dir, "flat-meta.json")) flat_sd = torch.load(flat_path) print(f"Loaded flat state dict from {flat_path}") output_sd = {} for flat_key, param_meta in flat_meta.items(): flat_param = flat_sd["model"][flat_key] assert ( sum(param_meta["numels"]) == flat_param.numel() ), f'flat {flat_key} {flat_param.numel()} vs {sum(param_meta["numels"])}' for name, shape, param in zip( param_meta["names"], param_meta["shapes"], flat_param.split(param_meta["numels"]) ): output_sd[name] = param.view(shape) torch.save(output_sd, output_path) print(f"Saved unflat state dict to {output_path}") if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("flat_dir") parser.add_argument("output_dir") parser.add_argument("part", type=int) args = parser.parse_args() convert(args.flat_dir, args.output_dir, args.part)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_vit/train.py
examples/tutorial/new_api/cifar_vit/train.py
import argparse import os from pathlib import Path import torch import torch.distributed as dist import torch.nn as nn import torchvision import torchvision.transforms as transforms from timm.models.vision_transformer import _cfg, _create_vision_transformer from torch.optim import Optimizer from torch.utils.data import DataLoader from tqdm import tqdm import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.booster.plugin.dp_plugin_base import DPPluginBase from colossalai.cluster import DistCoordinator from colossalai.nn.lr_scheduler import LinearWarmupLR from colossalai.nn.optimizer import HybridAdam # ============================== # Prepare Hyperparameters # ============================== NUM_EPOCHS = 60 WARMUP_EPOCHS = 5 LEARNING_RATE = 1e-3 def vit_cifar(**kwargs): pretrained_cfg = _cfg(num_classes=10, input_size=(3, 32, 32), crop_pct=1.0) model_kwargs = dict(patch_size=4, embed_dim=512, depth=6, num_heads=8, drop_rate=0.1, mlp_ratio=1.0, **kwargs) model = _create_vision_transformer("vit_cifar", pretrained_cfg=pretrained_cfg, **model_kwargs) return model def build_dataloader(batch_size: int, coordinator: DistCoordinator, plugin: DPPluginBase): # transform transform_train = transforms.Compose( [ transforms.RandomCrop(32, padding=4), transforms.RandomHorizontalFlip(), transforms.ToTensor(), transforms.Normalize((0.49139968, 0.48215827, 0.44653124), (0.24703233, 0.24348505, 0.26158768)), ] ) transform_test = transforms.Compose( [ transforms.Resize(32), transforms.ToTensor(), transforms.Normalize((0.49139968, 0.48215827, 0.44653124), (0.24703233, 0.24348505, 0.26158768)), ] ) # CIFAR-10 dataset data_path = os.environ.get("DATA", "./data") with coordinator.priority_execution(): train_dataset = torchvision.datasets.CIFAR10( root=data_path, train=True, transform=transform_train, download=True ) test_dataset = torchvision.datasets.CIFAR10( root=data_path, train=False, transform=transform_test, download=True ) # Data loader train_dataloader = plugin.prepare_dataloader(train_dataset, batch_size=batch_size, shuffle=True, drop_last=True) test_dataloader = plugin.prepare_dataloader(test_dataset, batch_size=batch_size, shuffle=False, drop_last=False) return train_dataloader, test_dataloader @torch.no_grad() def evaluate(model: nn.Module, test_dataloader: DataLoader, coordinator: DistCoordinator) -> float: model.eval() correct = torch.zeros(1, dtype=torch.int64, device=get_accelerator().get_current_device()) total = torch.zeros(1, dtype=torch.int64, device=get_accelerator().get_current_device()) for images, labels in test_dataloader: images = images.cuda() labels = labels.cuda() outputs = model(images) _, predicted = torch.max(outputs.data, 1) total += labels.size(0) correct += (predicted == labels).sum().item() dist.all_reduce(correct) dist.all_reduce(total) accuracy = correct.item() / total.item() if coordinator.is_master(): print(f"Accuracy of the model on the test images: {accuracy * 100:.2f} %") return accuracy def train_epoch( epoch: int, model: nn.Module, optimizer: Optimizer, criterion: nn.Module, train_dataloader: DataLoader, booster: Booster, coordinator: DistCoordinator, ): model.train() with tqdm(train_dataloader, desc=f"Epoch [{epoch + 1}/{NUM_EPOCHS}]", disable=not coordinator.is_master()) as pbar: for images, labels in pbar: images = images.cuda() labels = labels.cuda() # Forward pass outputs = model(images) loss = criterion(outputs, labels) # Backward and optimize booster.backward(loss, optimizer) optimizer.step() optimizer.zero_grad() # Print log info pbar.set_postfix({"loss": loss.item()}) def main(): # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() # FIXME(ver217): gemini is not supported resnet now parser.add_argument( "-p", "--plugin", type=str, default="torch_ddp", choices=["torch_ddp", "torch_ddp_fp16", "low_level_zero"], help="plugin to use", ) parser.add_argument("-r", "--resume", type=int, default=-1, help="resume from the epoch's checkpoint") parser.add_argument("-c", "--checkpoint", type=str, default="./checkpoint", help="checkpoint directory") parser.add_argument("-i", "--interval", type=int, default=5, help="interval of saving checkpoint") parser.add_argument( "--target_acc", type=float, default=None, help="target accuracy. Raise exception if not reached" ) args = parser.parse_args() # ============================== # Prepare Checkpoint Directory # ============================== if args.interval > 0: Path(args.checkpoint).mkdir(parents=True, exist_ok=True) # ============================== # Launch Distributed Environment # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # update the learning rate with linear scaling # old_gpu_num / old_lr = new_gpu_num / new_lr global LEARNING_RATE LEARNING_RATE *= coordinator.world_size # ============================== # Instantiate Plugin and Booster # ============================== booster_kwargs = {} if args.plugin == "torch_ddp_fp16": booster_kwargs["mixed_precision"] = "fp16" if args.plugin.startswith("torch_ddp"): plugin = TorchDDPPlugin() elif args.plugin == "gemini": plugin = GeminiPlugin(placement_policy="static", strict_ddp_mode=True, initial_scale=2**5) elif args.plugin == "low_level_zero": plugin = LowLevelZeroPlugin(initial_scale=2**5) booster = Booster(plugin=plugin, **booster_kwargs) # ============================== # Prepare Dataloader # ============================== train_dataloader, test_dataloader = build_dataloader(512, coordinator, plugin) # ==================================== # Prepare model, optimizer, criterion # ==================================== # resent50 model = torchvision.models.resnet18(num_classes=10) # Loss and optimizer criterion = nn.CrossEntropyLoss() optimizer = HybridAdam(model.parameters(), lr=LEARNING_RATE) # lr scheduler lr_scheduler = LinearWarmupLR(optimizer, NUM_EPOCHS, WARMUP_EPOCHS) # ============================== # Boost with ColossalAI # ============================== model, optimizer, criterion, train_dataloader, lr_scheduler = booster.boost( model, optimizer, criterion=criterion, dataloader=train_dataloader, lr_scheduler=lr_scheduler ) # ============================== # Resume from checkpoint # ============================== if args.resume >= 0: booster.load_model(model, f"{args.checkpoint}/model_{args.resume}.pth") booster.load_optimizer(optimizer, f"{args.checkpoint}/optimizer_{args.resume}.pth") booster.load_lr_scheduler(lr_scheduler, f"{args.checkpoint}/lr_scheduler_{args.resume}.pth") # ============================== # Train model # ============================== start_epoch = args.resume if args.resume >= 0 else 0 for epoch in range(start_epoch, NUM_EPOCHS): train_epoch(epoch, model, optimizer, criterion, train_dataloader, booster, coordinator) lr_scheduler.step() # save checkpoint if args.interval > 0 and (epoch + 1) % args.interval == 0: booster.save_model(model, f"{args.checkpoint}/model_{epoch + 1}.pth") booster.save_optimizer(optimizer, f"{args.checkpoint}/optimizer_{epoch + 1}.pth") booster.save_lr_scheduler(lr_scheduler, f"{args.checkpoint}/lr_scheduler_{epoch + 1}.pth") accuracy = evaluate(model, test_dataloader, coordinator) if args.target_acc is not None: assert accuracy >= args.target_acc, f"Accuracy {accuracy} is lower than target accuracy {args.target_acc}" if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/glue_bert/finetune.py
examples/tutorial/new_api/glue_bert/finetune.py
import argparse from typing import List, Union import datasets import torch import torch.distributed as dist import torch.nn as nn from data import GLUEDataBuilder from torch.optim import Optimizer from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoConfig, BertForSequenceClassification, get_linear_schedule_with_warmup import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.cluster import DistCoordinator from colossalai.nn.optimizer import HybridAdam # ============================== # Prepare Hyperparameters # ============================== NUM_EPOCHS = 1 BATCH_SIZE = 32 LEARNING_RATE = 2.4e-5 WEIGHT_DECAY = 0.01 WARMUP_FRACTION = 0.1 def move_to_cuda(batch): return {k: v.cuda() for k, v in batch.items()} @torch.no_grad() def evaluate( model: nn.Module, test_dataloader: Union[DataLoader, List[DataLoader]], num_labels: int, task_name: str, eval_splits: List[str], coordinator: DistCoordinator, ): metric = datasets.load_metric("glue", task_name, process_id=coordinator.rank, num_process=coordinator.world_size) model.eval() def evaluate_subset(dataloader: DataLoader): accum_loss = torch.zeros(1, device=get_accelerator().get_current_device()) for batch in dataloader: batch = move_to_cuda(batch) outputs = model(**batch) val_loss, logits = outputs[:2] accum_loss.add_(val_loss) if num_labels > 1: preds = torch.argmax(logits, axis=1) elif num_labels == 1: preds = logits.squeeze() labels = batch["labels"] metric.add_batch(predictions=preds, references=labels) results = metric.compute() dist.all_reduce(accum_loss.div_(len(dataloader))) if coordinator.is_master(): results["loss"] = accum_loss.item() / coordinator.world_size return results if isinstance(test_dataloader, DataLoader): return evaluate_subset(test_dataloader) else: assert len(test_dataloader) == len(eval_splits) final_results = {} for split, sub_loader in zip(eval_splits, test_dataloader): results = evaluate_subset(sub_loader) final_results.update({f"{k}_{split}": v for k, v in results.items()}) return final_results def train_epoch( epoch: int, model: nn.Module, optimizer: Optimizer, lr_scheduler, train_dataloader: DataLoader, booster: Booster, coordinator: DistCoordinator, ): model.train() with tqdm(train_dataloader, desc=f"Epoch [{epoch + 1}/{NUM_EPOCHS}]", disable=not coordinator.is_master()) as pbar: for batch in pbar: # Forward pass batch = move_to_cuda(batch) outputs = model(**batch) loss = outputs[0] # Backward and optimize booster.backward(loss, optimizer) optimizer.step() optimizer.zero_grad() lr_scheduler.step() # Print log info pbar.set_postfix({"loss": loss.item()}) def main(): # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument("-t", "--task", default="mrpc", help="GLUE task to run") parser.add_argument( "-p", "--plugin", type=str, default="torch_ddp", choices=["torch_ddp", "torch_ddp_fp16", "gemini", "low_level_zero"], help="plugin to use", ) parser.add_argument("--target_f1", type=float, default=None, help="target f1 score. Raise exception if not reached") args = parser.parse_args() # ============================== # Launch Distributed Environment # ============================== colossalai.launch_from_torch(seed=42) coordinator = DistCoordinator() # local_batch_size = BATCH_SIZE // coordinator.world_size lr = LEARNING_RATE * coordinator.world_size model_name = "bert-base-uncased" # ============================== # Instantiate Plugin and Booster # ============================== booster_kwargs = {} if args.plugin == "torch_ddp_fp16": booster_kwargs["mixed_precision"] = "fp16" if args.plugin.startswith("torch_ddp"): plugin = TorchDDPPlugin() elif args.plugin == "gemini": plugin = GeminiPlugin(placement_policy="static", strict_ddp_mode=True, initial_scale=2**5) elif args.plugin == "low_level_zero": plugin = LowLevelZeroPlugin(initial_scale=2**5) booster = Booster(plugin=plugin, **booster_kwargs) # ============================== # Prepare Dataloader # ============================== data_builder = GLUEDataBuilder( model_name, plugin, args.task, train_batch_size=BATCH_SIZE, eval_batch_size=BATCH_SIZE ) train_dataloader = data_builder.train_dataloader() test_dataloader = data_builder.test_dataloader() # ==================================== # Prepare model, optimizer # ==================================== # bert pretrained model config = AutoConfig.from_pretrained(model_name, num_labels=data_builder.num_labels) model = BertForSequenceClassification.from_pretrained(model_name, config=config) # optimizer no_decay = ["bias", "LayerNorm.weight"] optimizer_grouped_parameters = [ { "params": [p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay)], "weight_decay": WEIGHT_DECAY, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], "weight_decay": 0.0, }, ] optimizer = HybridAdam(optimizer_grouped_parameters, lr=lr, eps=1e-8) # lr scheduler total_steps = len(train_dataloader) * NUM_EPOCHS num_warmup_steps = int(WARMUP_FRACTION * total_steps) lr_scheduler = get_linear_schedule_with_warmup( optimizer, num_warmup_steps=num_warmup_steps, num_training_steps=total_steps, ) # ============================== # Boost with ColossalAI # ============================== model, optimizer, _, _, lr_scheduler = booster.boost(model, optimizer, lr_scheduler=lr_scheduler) # ============================== # Train model # ============================== for epoch in range(NUM_EPOCHS): train_epoch(epoch, model, optimizer, lr_scheduler, train_dataloader, booster, coordinator) results = evaluate( model, test_dataloader, data_builder.num_labels, args.task, data_builder.eval_splits, coordinator ) if coordinator.is_master(): print(results) if args.target_f1 is not None and "f1" in results: assert results["f1"] >= args.target_f1, f'f1 score {results["f1"]} is lower than target {args.target_f1}' if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/glue_bert/data.py
examples/tutorial/new_api/glue_bert/data.py
import datasets from transformers import AutoTokenizer, PreTrainedTokenizer from colossalai.booster.plugin.dp_plugin_base import DPPluginBase class GLUEDataBuilder: task_text_field_map = { "cola": ["sentence"], "sst2": ["sentence"], "mrpc": ["sentence1", "sentence2"], "qqp": ["question1", "question2"], "stsb": ["sentence1", "sentence2"], "mnli": ["premise", "hypothesis"], "qnli": ["question", "sentence"], "rte": ["sentence1", "sentence2"], "wnli": ["sentence1", "sentence2"], "ax": ["premise", "hypothesis"], } glue_task_num_labels = { "cola": 2, "sst2": 2, "mrpc": 2, "qqp": 2, "stsb": 1, "mnli": 3, "qnli": 2, "rte": 2, "wnli": 2, "ax": 3, } loader_columns = [ "datasets_idx", "input_ids", "token_type_ids", "attention_mask", "start_positions", "end_positions", "labels", ] def __init__( self, model_name_or_path: str, plugin: DPPluginBase, task_name: str = "mrpc", max_seq_length: int = 128, train_batch_size: int = 32, eval_batch_size: int = 32, **kwargs, ): super().__init__() self.model_name_or_path = model_name_or_path self.task_name = task_name self.max_seq_length = max_seq_length self.train_batch_size = train_batch_size self.eval_batch_size = eval_batch_size self.plugin = plugin self.text_fields = self.task_text_field_map[task_name] self.num_labels = self.glue_task_num_labels[task_name] self.tokenizer: PreTrainedTokenizer = AutoTokenizer.from_pretrained(self.model_name_or_path, use_fast=True) self.setup() def setup(self): self.dataset = datasets.load_dataset("glue", self.task_name) for split in self.dataset.keys(): self.dataset[split] = self.dataset[split].map( self.convert_to_features, batched=True, remove_columns=["label"], ) self.columns = [c for c in self.dataset[split].column_names if c in self.loader_columns] self.dataset[split].set_format(type="torch", columns=self.columns) self.eval_splits = [x for x in self.dataset.keys() if "validation" in x] def prepare_data(self): datasets.load_dataset("glue", self.task_name) AutoTokenizer.from_pretrained(self.model_name_or_path, use_fast=True) def train_dataloader(self): return self.plugin.prepare_dataloader( self.dataset["train"], batch_size=self.train_batch_size, shuffle=True, drop_last=True ) def val_dataloader(self): if len(self.eval_splits) == 1: return self.plugin.prepare_dataloader(self.dataset["validation"], batch_size=self.eval_batch_size) elif len(self.eval_splits) > 1: return [ self.plugin.prepare_dataloader(self.dataset[x], batch_size=self.eval_batch_size) for x in self.eval_splits ] def test_dataloader(self): if len(self.eval_splits) == 1: return self.plugin.prepare_dataloader(self.dataset["test"], batch_size=self.eval_batch_size) elif len(self.eval_splits) > 1: return [ self.plugin.prepare_dataloader(self.dataset[x], batch_size=self.eval_batch_size) for x in self.eval_splits ] def convert_to_features(self, example_batch): # Either encode single sentence or sentence pairs if len(self.text_fields) > 1: texts_or_text_pairs = list(zip(example_batch[self.text_fields[0]], example_batch[self.text_fields[1]])) else: texts_or_text_pairs = example_batch[self.text_fields[0]] # Tokenize the text/text pairs features = self.tokenizer.batch_encode_plus( texts_or_text_pairs, max_length=self.max_seq_length, padding="max_length", truncation=True ) # Rename label to labels to make it easier to pass to model forward features["labels"] = example_batch["label"] return features
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_resnet/train.py
examples/tutorial/new_api/cifar_resnet/train.py
import argparse import os from pathlib import Path import torch import torch.distributed as dist import torch.nn as nn import torchvision import torchvision.transforms as transforms from torch.optim import Optimizer from torch.optim.lr_scheduler import MultiStepLR from torch.utils.data import DataLoader from tqdm import tqdm import colossalai from colossalai.accelerator import get_accelerator from colossalai.booster import Booster from colossalai.booster.plugin import GeminiPlugin, LowLevelZeroPlugin, TorchDDPPlugin from colossalai.booster.plugin.dp_plugin_base import DPPluginBase from colossalai.cluster import DistCoordinator from colossalai.nn.optimizer import HybridAdam # ============================== # Prepare Hyperparameters # ============================== NUM_EPOCHS = 80 LEARNING_RATE = 1e-3 def build_dataloader(batch_size: int, coordinator: DistCoordinator, plugin: DPPluginBase): # transform transform_train = transforms.Compose( [transforms.Pad(4), transforms.RandomHorizontalFlip(), transforms.RandomCrop(32), transforms.ToTensor()] ) transform_test = transforms.ToTensor() # CIFAR-10 dataset data_path = os.environ.get("DATA", "./data") with coordinator.priority_execution(): train_dataset = torchvision.datasets.CIFAR10( root=data_path, train=True, transform=transform_train, download=True ) test_dataset = torchvision.datasets.CIFAR10( root=data_path, train=False, transform=transform_test, download=True ) # Data loader train_dataloader = plugin.prepare_dataloader(train_dataset, batch_size=batch_size, shuffle=True, drop_last=True) test_dataloader = plugin.prepare_dataloader(test_dataset, batch_size=batch_size, shuffle=False, drop_last=False) return train_dataloader, test_dataloader @torch.no_grad() def evaluate(model: nn.Module, test_dataloader: DataLoader, coordinator: DistCoordinator) -> float: model.eval() correct = torch.zeros(1, dtype=torch.int64, device=get_accelerator().get_current_device()) total = torch.zeros(1, dtype=torch.int64, device=get_accelerator().get_current_device()) for images, labels in test_dataloader: images = images.cuda() labels = labels.cuda() outputs = model(images) _, predicted = torch.max(outputs.data, 1) total += labels.size(0) correct += (predicted == labels).sum().item() dist.all_reduce(correct) dist.all_reduce(total) accuracy = correct.item() / total.item() if coordinator.is_master(): print(f"Accuracy of the model on the test images: {accuracy * 100:.2f} %") return accuracy def train_epoch( epoch: int, model: nn.Module, optimizer: Optimizer, criterion: nn.Module, train_dataloader: DataLoader, booster: Booster, coordinator: DistCoordinator, ): model.train() with tqdm(train_dataloader, desc=f"Epoch [{epoch + 1}/{NUM_EPOCHS}]", disable=not coordinator.is_master()) as pbar: for images, labels in pbar: images = images.cuda() labels = labels.cuda() # Forward pass outputs = model(images) loss = criterion(outputs, labels) # Backward and optimize booster.backward(loss, optimizer) optimizer.step() optimizer.zero_grad() # Print log info pbar.set_postfix({"loss": loss.item()}) def main(): # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() # FIXME(ver217): gemini is not supported resnet now parser.add_argument( "-p", "--plugin", type=str, default="torch_ddp", choices=["torch_ddp", "torch_ddp_fp16", "low_level_zero"], help="plugin to use", ) parser.add_argument("-r", "--resume", type=int, default=-1, help="resume from the epoch's checkpoint") parser.add_argument("-c", "--checkpoint", type=str, default="./checkpoint", help="checkpoint directory") parser.add_argument("-i", "--interval", type=int, default=5, help="interval of saving checkpoint") parser.add_argument( "--target_acc", type=float, default=None, help="target accuracy. Raise exception if not reached" ) args = parser.parse_args() # ============================== # Prepare Checkpoint Directory # ============================== if args.interval > 0: Path(args.checkpoint).mkdir(parents=True, exist_ok=True) # ============================== # Launch Distributed Environment # ============================== colossalai.launch_from_torch() coordinator = DistCoordinator() # update the learning rate with linear scaling # old_gpu_num / old_lr = new_gpu_num / new_lr global LEARNING_RATE LEARNING_RATE *= coordinator.world_size # ============================== # Instantiate Plugin and Booster # ============================== booster_kwargs = {} if args.plugin == "torch_ddp_fp16": booster_kwargs["mixed_precision"] = "fp16" if args.plugin.startswith("torch_ddp"): plugin = TorchDDPPlugin() elif args.plugin == "gemini": plugin = GeminiPlugin(placement_policy="static", strict_ddp_mode=True, initial_scale=2**5) elif args.plugin == "low_level_zero": plugin = LowLevelZeroPlugin(initial_scale=2**5) booster = Booster(plugin=plugin, **booster_kwargs) # ============================== # Prepare Dataloader # ============================== train_dataloader, test_dataloader = build_dataloader(100, coordinator, plugin) # ==================================== # Prepare model, optimizer, criterion # ==================================== # resent50 model = torchvision.models.resnet18(num_classes=10) # Loss and optimizer criterion = nn.CrossEntropyLoss() optimizer = HybridAdam(model.parameters(), lr=LEARNING_RATE) # lr scheduler lr_scheduler = MultiStepLR(optimizer, milestones=[20, 40, 60, 80], gamma=1 / 3) # ============================== # Boost with ColossalAI # ============================== model, optimizer, criterion, _, lr_scheduler = booster.boost( model, optimizer, criterion=criterion, lr_scheduler=lr_scheduler ) # ============================== # Resume from checkpoint # ============================== if args.resume >= 0: booster.load_model(model, f"{args.checkpoint}/model_{args.resume}.pth") booster.load_optimizer(optimizer, f"{args.checkpoint}/optimizer_{args.resume}.pth") booster.load_lr_scheduler(lr_scheduler, f"{args.checkpoint}/lr_scheduler_{args.resume}.pth") # ============================== # Train model # ============================== start_epoch = args.resume if args.resume >= 0 else 0 for epoch in range(start_epoch, NUM_EPOCHS): train_epoch(epoch, model, optimizer, criterion, train_dataloader, booster, coordinator) lr_scheduler.step() # save checkpoint if args.interval > 0 and (epoch + 1) % args.interval == 0: booster.save_model(model, f"{args.checkpoint}/model_{epoch + 1}.pth") booster.save_optimizer(optimizer, f"{args.checkpoint}/optimizer_{epoch + 1}.pth") booster.save_lr_scheduler(lr_scheduler, f"{args.checkpoint}/lr_scheduler_{epoch + 1}.pth") accuracy = evaluate(model, test_dataloader, coordinator) if args.target_acc is not None: assert accuracy >= args.target_acc, f"Accuracy {accuracy} is lower than target accuracy {args.target_acc}" if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/new_api/cifar_resnet/eval.py
examples/tutorial/new_api/cifar_resnet/eval.py
import argparse import torch import torchvision import torchvision.transforms as transforms # ============================== # Parse Arguments # ============================== parser = argparse.ArgumentParser() parser.add_argument("-e", "--epoch", type=int, default=80, help="resume from the epoch's checkpoint") parser.add_argument("-c", "--checkpoint", type=str, default="./checkpoint", help="checkpoint directory") args = parser.parse_args() # ============================== # Prepare Test Dataset # ============================== # CIFAR-10 dataset test_dataset = torchvision.datasets.CIFAR10(root="./data/", train=False, transform=transforms.ToTensor()) # Data loader test_loader = torch.utils.data.DataLoader(dataset=test_dataset, batch_size=128, shuffle=False) # ============================== # Load Model # ============================== model = torchvision.models.resnet18(num_classes=10).cuda() state_dict = torch.load(f"{args.checkpoint}/model_{args.epoch}.pth") model.load_state_dict(state_dict) # ============================== # Run Evaluation # ============================== model.eval() with torch.no_grad(): correct = 0 total = 0 for images, labels in test_loader: images = images.cuda() labels = labels.cuda() outputs = model(images) _, predicted = torch.max(outputs.data, 1) total += labels.size(0) correct += (predicted == labels).sum().item() print("Accuracy of the model on the test images: {} %".format(100 * correct / total))
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/train.py
examples/tutorial/sequence_parallel/train.py
import argparse import torch from data.bert_helper import SequenceParallelDataIterator, get_batch_for_sequence_parallel from data.dummy_dataloader import DummyDataloader from loss_func.bert_loss import BertLoss from lr_scheduler import AnnealingLR from model.bert import BertForPretrain, build_pipeline_bert import colossalai from colossalai.legacy.amp import AMP_TYPE from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.utils import is_using_pp from colossalai.logging import get_dist_logger from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from colossalai.nn.optimizer import FusedAdam from colossalai.utils import MultiTimer def process_batch_data(batch_data): tokens, types, sentence_order, loss_mask, lm_labels, padding_mask = batch_data if gpc.is_first_rank(ParallelMode.PIPELINE): data = dict(input_ids=tokens, attention_masks=padding_mask, tokentype_ids=types, lm_labels=lm_labels) else: data = dict(attention_masks=padding_mask, tokentype_ids=types, lm_labels=lm_labels) label = dict(loss_mask=loss_mask, sentence_order=sentence_order) return data, label def parse_args(): parser = argparse.ArgumentParser() parser.add_argument("-s", "--synthetic", action="store_true", help="whether use synthetic data") return parser.parse_args() def pipeline_data_process_func(stage_output, micro_batch_data): tokens, types, sentence_order, loss_mask, lm_labels, padding_mask = micro_batch_data if gpc.is_first_rank(ParallelMode.PIPELINE): data = (tokens, padding_mask, types, lm_labels) label = (loss_mask, sentence_order) else: data = (stage_output, padding_mask, types, lm_labels) label = (loss_mask, sentence_order) return data, label def main(): # initialize parse_args() colossalai.legacy.launch_from_torch(config="./config.py", seed=1234, backend="nccl") logger = get_dist_logger() # build synthetic dataloader BATCH_SIZE_PER_GPUS = gpc.config.GLOBAL_BATCH_SIZE // gpc.get_world_size(ParallelMode.DATA) VOCAB_SIZE = 30528 trainloader = DummyDataloader( batch_size=BATCH_SIZE_PER_GPUS, vocab_size=VOCAB_SIZE, seq_length=gpc.config.SEQ_LENGTH ) validloader = DummyDataloader( batch_size=BATCH_SIZE_PER_GPUS, vocab_size=VOCAB_SIZE, seq_length=gpc.config.SEQ_LENGTH ) logger.info("Dataloaders are built", ranks=[0]) # build model if hasattr(gpc.config, "fp16") and gpc.config.fp16.get("mode") == AMP_TYPE.NAIVE: is_naive_fp16 = True else: is_naive_fp16 = False use_pipeline = is_using_pp() kwargs = dict( vocab_size=VOCAB_SIZE, hidden_size=gpc.config.HIDDEN_SIZE, max_sequence_length=gpc.config.SEQ_LENGTH, num_attention_heads=gpc.config.NUM_ATTENTION_HEADS, convert_fp16_to_fp32_in_softmax=True, is_naive_fp16=is_naive_fp16, add_binary_head=gpc.config.ADD_BINARY_HEAD, ) if use_pipeline: model = build_pipeline_bert(num_layers=gpc.config.DEPTH, num_chunks=1, **kwargs) else: model = BertForPretrain(num_layers=gpc.config.DEPTH, **kwargs) model = model.half() model.reset_parameters() logger.info(f"Model is built with softmax in fp32 = {is_naive_fp16}", ranks=[0]) total_numel = 0 for p in model.parameters(): total_numel += p.numel() logger.info(f"This model has {total_numel} parameters") # build criterion criterion = BertLoss() logger.info("Criterion is built", ranks=[0]) # layernorm and bias has no weight decay weight_decay_params = {"params": []} no_weight_decay_params = {"params": [], "weight_decay": 0.0} for module_ in model.modules(): if isinstance(module_, LayerNorm): no_weight_decay_params["params"].extend([p for p in list(module_._parameters.values()) if p is not None]) else: weight_decay_params["params"].extend( [p for n, p in list(module_._parameters.items()) if p is not None and n != "bias"] ) no_weight_decay_params["params"].extend( [p for n, p in list(module_._parameters.items()) if p is not None and n == "bias"] ) logger.info( f"without weight decay param: {len(no_weight_decay_params['params'])}, with weight decay param: {len(weight_decay_params['params'])}" ) # optimizer optimizer = FusedAdam( (weight_decay_params, no_weight_decay_params), lr=gpc.config.LR, weight_decay=gpc.config.WEIGHT_DECAY ) logger.info("Optimizer is built", ranks=[0]) # lr scheduler # follow Megatron-LM setting warmup_steps = int(gpc.config.DECAY_ITERS * gpc.config.WARMUP_FRACTION) lr_scheduler = AnnealingLR( optimizer=optimizer, max_lr=gpc.config.LR, min_lr=gpc.config.MIN_LR, warmup_steps=warmup_steps, decay_steps=gpc.config.DECAY_ITERS, decay_style="linear", ) logger.info(f"LR Scheduler is built with {warmup_steps} warmup steps and {gpc.config.DECAY_ITERS} decay steps") # # init engine, *dummy = colossalai.legacy.initialize(model, optimizer, criterion, verbose=True) # build timer timer = MultiTimer() # build loss tracker accumulated_train_loss = torch.zeros(1, dtype=torch.float32).cuda() accumulated_eval_loss = torch.zeros(1, dtype=torch.float32).cuda() # build data iters for pipeline parallel if use_pipeline: train_data_iter = SequenceParallelDataIterator(trainloader) valid_data_iter = SequenceParallelDataIterator(validloader) engine.schedule.data_process_func = pipeline_data_process_func logger.info("start training") for step in range(1, gpc.config.TRAIN_ITERS + 1): timer.start("train-iterations") engine.train() if use_pipeline: engine.zero_grad() _, _, train_loss = engine.execute_schedule(train_data_iter, return_output_label=False) engine.step() else: tokens, types, sentence_order, loss_mask, lm_labels, padding_mask = get_batch_for_sequence_parallel( trainloader ) engine.zero_grad() lm_loss, sop_output = engine(tokens, padding_mask, types, lm_labels) train_loss = engine.criterion(lm_loss, sop_output, loss_mask, sentence_order) engine.backward(train_loss) engine.step() timer.stop("train-iterations", keep_in_history=True) if not gpc.is_initialized(ParallelMode.PIPELINE) or gpc.is_last_rank(ParallelMode.PIPELINE): accumulated_train_loss += train_loss lr_scheduler.step() if step % gpc.config.EVAL_INTERVAL == 0: engine.eval() for j in range(gpc.config.EVAL_ITERS): with torch.no_grad(): if use_pipeline: _, _, eval_loss = engine.execute_schedule( valid_data_iter, forward_only=True, return_output_label=False ) else: ( tokens, types, sentence_order, loss_mask, lm_labels, padding_mask, ) = get_batch_for_sequence_parallel(validloader) lm_loss, sop_output = engine(tokens, padding_mask, types, lm_labels) eval_loss = engine.criterion(lm_loss, sop_output, loss_mask, sentence_order) if not gpc.is_initialized(ParallelMode.PIPELINE) or gpc.is_last_rank(ParallelMode.PIPELINE): accumulated_eval_loss += eval_loss if not gpc.is_initialized(ParallelMode.PIPELINE) or gpc.is_last_rank(ParallelMode.PIPELINE): accumulated_eval_loss /= gpc.config.EVAL_ITERS accumulated_train_loss /= gpc.config.EVAL_INTERVAL timer_string = [] for n, t in timer: timer_string.append(f"{n}: {t.get_history_mean()*1000:.5f}") timer_string = " | ".join(timer_string) lr = list(engine.optimizer.param_groups)[0]["lr"] loss_scale = engine.optimizer.optim.loss_scale.item() if gpc.is_initialized(ParallelMode.PIPELINE): ranks = [gpc.get_ranks_in_group(ParallelMode.PIPELINE)[-1]] else: ranks = [0] logger.info( f"Step {step} / {gpc.config.TRAIN_ITERS} | Train Loss: {accumulated_train_loss.item():.5g} " + f"| Eval Loss: {accumulated_eval_loss.item():.5g} " + f"| Loss Scale: {loss_scale}" + f"| Learning rate: {lr} | " + timer_string, ranks=ranks, ) for n, t in timer: t.reset() accumulated_eval_loss.zero_() accumulated_train_loss.zero_() if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/config.py
examples/tutorial/sequence_parallel/config.py
from colossalai.legacy.amp import AMP_TYPE # hyper-parameters TRAIN_ITERS = 10 DECAY_ITERS = 4 WARMUP_FRACTION = 0.01 GLOBAL_BATCH_SIZE = 32 # dp world size * sentences per GPU EVAL_ITERS = 10 EVAL_INTERVAL = 10 LR = 0.0001 MIN_LR = 1e-05 WEIGHT_DECAY = 0.01 SEQ_LENGTH = 128 # BERT config DEPTH = 4 NUM_ATTENTION_HEADS = 4 HIDDEN_SIZE = 128 # model config ADD_BINARY_HEAD = False # random seed SEED = 1234 # pipeline config # only enabled when pipeline > 1 NUM_MICRO_BATCHES = 4 # colossalai config parallel = dict(pipeline=1, tensor=dict(size=2, mode="sequence")) fp16 = dict(mode=AMP_TYPE.NAIVE, verbose=True) gradient_handler = [dict(type="SequenceParallelGradientHandler")]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/bert_loss.py
examples/tutorial/sequence_parallel/loss_func/bert_loss.py
import torch import torch.nn as nn import torch.nn.functional as F from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc class BertLoss(nn.Module): def forward(self, lm_loss, sop_logits, loss_mask, sentence_order): lm_loss_ = lm_loss.float() loss_mask = loss_mask.float() loss_mask_sum = loss_mask.sum() lm_loss = torch.sum(lm_loss_.view(-1) * loss_mask.reshape(-1)) lm_loss /= loss_mask_sum torch.distributed.all_reduce(lm_loss, group=gpc.get_group(ParallelMode.SEQUENCE)) if sop_logits is not None: sop_loss = F.cross_entropy(sop_logits.view(-1, 2).float(), sentence_order.view(-1), ignore_index=-1) sop_loss = sop_loss.float() loss = lm_loss + sop_loss * gpc.get_world_size(ParallelMode.SEQUENCE) else: sop_loss = None loss = lm_loss return loss
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/cross_entropy.py
examples/tutorial/sequence_parallel/loss_func/cross_entropy.py
import torch from torch.cuda.amp import custom_bwd, custom_fwd class _VocabCrossEntropy(torch.autograd.Function): @staticmethod @custom_fwd def forward(ctx, vocab_parallel_logits, target): # Maximum value along vocab dimension across all GPUs. logits_max = torch.max(vocab_parallel_logits, dim=-1)[0] # Subtract the maximum value. vocab_parallel_logits.sub_(logits_max.unsqueeze(dim=-1)) # Create a mask of valid vocab ids (1 means it needs to be masked). target_mask = target < 0 masked_target = target.clone() masked_target[target_mask] = 0 # Get predicted-logits = logits[target]. # For Simplicity, we convert logits to a 2-D tensor with size # [*, partition-vocab-size] and target to a 1-D tensor of size [*]. logits_2d = vocab_parallel_logits.view(-1, vocab_parallel_logits.size(-1)) masked_target_1d = masked_target.view(-1) arange_1d = torch.arange(start=0, end=logits_2d.size()[0], device=logits_2d.device) predicted_logits_1d = logits_2d[arange_1d, masked_target_1d] predicted_logits_1d = predicted_logits_1d.clone().contiguous() predicted_logits = predicted_logits_1d.view_as(target) predicted_logits[target_mask] = 0.0 # Sum of exponential of logits along vocab dimension across all GPUs. exp_logits = vocab_parallel_logits torch.exp(vocab_parallel_logits, out=exp_logits) sum_exp_logits = exp_logits.sum(dim=-1) # Loss = log(sum(exp(logits))) - predicted-logit. loss = torch.log(sum_exp_logits) - predicted_logits # Store softmax, target-mask and masked-target for backward pass. exp_logits.div_(sum_exp_logits.unsqueeze(dim=-1)) ctx.save_for_backward(exp_logits, target_mask, masked_target_1d) return loss @staticmethod @custom_bwd def backward(ctx, grad_output): # Retreive tensors from the forward path. softmax, target_mask, masked_target_1d = ctx.saved_tensors # All the inputs have softmax as their gradient. grad_input = softmax # For simplicity, work with the 2D gradient. partition_vocab_size = softmax.size()[-1] grad_2d = grad_input.view(-1, partition_vocab_size) # Add the gradient from matching classes. arange_1d = torch.arange(start=0, end=grad_2d.size()[0], device=grad_2d.device) grad_2d[arange_1d, masked_target_1d] -= 1.0 - target_mask.view(-1).float() # Finally elementwise multiplication with the output gradients. grad_input.mul_(grad_output.unsqueeze(dim=-1)) return grad_input, None def vocab_cross_entropy(vocab_logits, target): """helper function for the cross entropy.""" return _VocabCrossEntropy.apply(vocab_logits, target)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/utils.py
examples/tutorial/sequence_parallel/loss_func/utils.py
import torch def ensure_divisibility(numerator, denominator): """Ensure that numerator is divisible by the denominator.""" assert numerator % denominator == 0, "{} is not divisible by {}".format(numerator, denominator) def divide(numerator, denominator): """Ensure that numerator is divisible by the denominator and return the division value.""" ensure_divisibility(numerator, denominator) return numerator // denominator def split_tensor_along_last_dim(tensor, num_partitions, contiguous_split_chunks=False): """Split a tensor along its last dimension. Arguments: tensor: input tensor. num_partitions: number of partitions to split the tensor contiguous_split_chunks: If True, make each chunk contiguous in memory. """ # Get the size and dimension. last_dim = tensor.dim() - 1 last_dim_size = divide(tensor.size()[last_dim], num_partitions) # Split. tensor_list = torch.split(tensor, last_dim_size, dim=last_dim) # Note: torch.split does not create contiguous tensors by default. if contiguous_split_chunks: return tuple(chunk.contiguous() for chunk in tensor_list) return tensor_list class VocabUtility: """Split the vocabulary into `world_size` chunks amd return the first and last index of the vocabulary belonging to the `rank` partition: Note that indices in [fist, last)""" @staticmethod def vocab_range_from_per_partition_vocab_size(per_partition_vocab_size, rank, world_size): index_f = rank * per_partition_vocab_size index_l = index_f + per_partition_vocab_size return index_f, index_l @staticmethod def vocab_range_from_global_vocab_size(global_vocab_size, rank, world_size): per_partition_vocab_size = divide(global_vocab_size, world_size) return VocabUtility.vocab_range_from_per_partition_vocab_size(per_partition_vocab_size, rank, world_size)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/loss_func/__init__.py
examples/tutorial/sequence_parallel/loss_func/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/bert.py
examples/tutorial/sequence_parallel/model/bert.py
import inspect import torch import torch.nn as nn from colossalai.legacy.context import ParallelMode from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.legacy.nn.layer.wrapper import PipelineSharedModuleWrapper from colossalai.legacy.pipeline.utils import partition_uniform from colossalai.logging import get_dist_logger from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from .layers import BertDualHead, BertLayer, Embedding, PreProcessor, VocabEmbedding from .layers.init_method import init_normal, output_init_normal class BertForPretrain(nn.Module): def __init__( self, vocab_size, hidden_size, max_sequence_length, num_attention_heads, num_layers, add_binary_head, is_naive_fp16, num_tokentypes=2, dropout_prob=0.1, mlp_ratio=4, init_std=0.02, convert_fp16_to_fp32_in_softmax=False, ): super().__init__() self.seq_parallel_size = gpc.get_world_size(ParallelMode.SEQUENCE) assert ( max_sequence_length % self.seq_parallel_size == 0 ), "sequence length is not divisible by the sequence parallel size" self.sub_seq_length = max_sequence_length // self.seq_parallel_size self.init_std = init_std self.num_layers = num_layers if not add_binary_head: num_tokentypes = 0 self.preprocessor = PreProcessor(self.sub_seq_length) self.embedding = Embedding( hidden_size=hidden_size, vocab_size=vocab_size, max_sequence_length=max_sequence_length, embedding_dropout_prob=dropout_prob, num_tokentypes=num_tokentypes, ) self.bert_layers = nn.ModuleList() for i in range(num_layers): bert_layer = BertLayer( layer_number=i + 1, hidden_size=hidden_size, num_attention_heads=num_attention_heads, attention_dropout=dropout_prob, mlp_ratio=mlp_ratio, hidden_dropout=dropout_prob, convert_fp16_to_fp32_in_softmax=convert_fp16_to_fp32_in_softmax, is_naive_fp16=is_naive_fp16, ) self.bert_layers.append(bert_layer) self.layer_norm = LayerNorm(hidden_size) self.head = BertDualHead( hidden_size, self.embedding.word_embedding_weight.size(0), add_binary_head=add_binary_head ) self.reset_parameters() def _init_normal(self, tensor): init_normal(tensor, sigma=self.init_std) def _output_init_normal(self, tensor): output_init_normal(tensor, sigma=self.init_std, num_layers=self.num_layers) def reset_parameters(self): # initialize embedding self._init_normal(self.embedding.word_embedding_weight) self._init_normal(self.embedding.position_embeddings.weight) if self.embedding.tokentype_embeddings: self._init_normal(self.embedding.tokentype_embeddings.weight) # initialize bert layer for layer in self.bert_layers: # initialize self attention self._init_normal(layer.self_attention.query_key_value.weight) self._output_init_normal(layer.self_attention.dense.weight) self._init_normal(layer.mlp.dense_h_to_4h.weight) self._output_init_normal(layer.mlp.dense_4h_to_h.weight) # initializer head self._init_normal(self.head.lm_head.dense.weight) if self.head.binary_head is not None: self._init_normal(self.head.binary_head.pooler.dense.weight) self._init_normal(self.head.binary_head.dense.weight) def forward(self, input_ids, attention_masks, tokentype_ids, lm_labels): # inputs of the forward function # input_ids: [batch_size, sub_seq_len] # attention_mask: [batch_size, seq_len] # tokentype_ids: [batch_size, sub_seq_len] # outputs of preprocessor # pos_ids: [batch_size, sub_seq_len] # attention_masks: [batch_size, 1, sub_seq_len, seq_len] pos_ids, attention_masks = self.preprocessor(input_ids, attention_masks) hidden_states = self.embedding(input_ids, pos_ids, tokentype_ids) # hidden_states shape change: # [batch_size, sub_seq_len, hidden_size] -> [sub_seq_len, batch_size, hidden_size] hidden_states = hidden_states.transpose(0, 1).contiguous() for idx, layer in enumerate(self.bert_layers): hidden_states = layer(hidden_states, attention_masks) hidden_states = hidden_states.transpose(0, 1).contiguous() output = self.layer_norm(hidden_states) # hidden_states: [sub_seq_len, batch_size, hidden_size] # word_embedding: [vocab_size, hidden_size] return self.head(output, self.embedding.word_embedding_weight, lm_labels) class PipelineBertForPretrain(nn.Module): def __init__( self, vocab_size, hidden_size, max_sequence_length, num_attention_heads, num_layers, add_binary_head, is_naive_fp16, num_tokentypes=2, dropout_prob=0.1, mlp_ratio=4, init_std=0.02, convert_fp16_to_fp32_in_softmax=False, first_stage=True, last_stage=True, start_idx=None, end_idx=None, ): super().__init__() self.seq_parallel_size = gpc.get_world_size(ParallelMode.SEQUENCE) assert ( max_sequence_length % self.seq_parallel_size == 0 ), "sequence length is not divisible by the sequence parallel size" self.sub_seq_length = max_sequence_length // self.seq_parallel_size self.init_std = init_std self.num_layers = num_layers if not add_binary_head: num_tokentypes = 0 self.first_stage = first_stage self.last_stage = last_stage self.preprocessor = PreProcessor(self.sub_seq_length) if self.first_stage: self.embedding = Embedding( hidden_size=hidden_size, vocab_size=vocab_size, max_sequence_length=max_sequence_length, embedding_dropout_prob=dropout_prob, num_tokentypes=num_tokentypes, ) # transformer layers self.bert_layers = nn.ModuleList() if start_idx is None and end_idx is None: start_idx = 0 end_idx = num_layers for i in range(start_idx, end_idx): bert_layer = BertLayer( layer_number=i + 1, hidden_size=hidden_size, num_attention_heads=num_attention_heads, attention_dropout=dropout_prob, mlp_ratio=mlp_ratio, hidden_dropout=dropout_prob, convert_fp16_to_fp32_in_softmax=convert_fp16_to_fp32_in_softmax, is_naive_fp16=is_naive_fp16, ) self.bert_layers.append(bert_layer) if self.last_stage: self.word_embeddings = VocabEmbedding(vocab_size, hidden_size) self.layer_norm = LayerNorm(hidden_size) self.head = BertDualHead(hidden_size, vocab_size, add_binary_head=add_binary_head) self.reset_parameters() def _init_normal(self, tensor): init_normal(tensor, sigma=self.init_std) def _output_init_normal(self, tensor): output_init_normal(tensor, sigma=self.init_std, num_layers=self.num_layers) def reset_parameters(self): # initialize embedding if self.first_stage: self._init_normal(self.embedding.word_embedding_weight) self._init_normal(self.embedding.position_embeddings.weight) if self.embedding.tokentype_embeddings: self._init_normal(self.embedding.tokentype_embeddings.weight) # initialize bert layer for layer in self.bert_layers: # initialize self attention self._init_normal(layer.self_attention.query_key_value.weight) self._output_init_normal(layer.self_attention.dense.weight) self._init_normal(layer.mlp.dense_h_to_4h.weight) self._output_init_normal(layer.mlp.dense_4h_to_h.weight) # initializer head if self.last_stage: self._init_normal(self.head.lm_head.dense.weight) if self.head.binary_head is not None: self._init_normal(self.head.binary_head.pooler.dense.weight) self._init_normal(self.head.binary_head.dense.weight) def forward(self, input_ids, attention_masks, tokentype_ids, lm_labels): # inputs of the forward function # input_ids: [batch_size, sub_seq_len] # attention_mask: [batch_size, seq_len] # tokentype_ids: [batch_size, sub_seq_len] # outputs of preprocessor # pos_ids: [batch_size, sub_seq_len] # attention_masks: [batch_size, 1, sub_seq_len, seq_len] if self.first_stage: pos_ids, attention_masks = self.preprocessor(input_ids, attention_masks) else: _, attention_masks = self.preprocessor(None, attention_masks) if self.first_stage: hidden_states = self.embedding(input_ids, pos_ids, tokentype_ids) hidden_states = hidden_states.transpose(0, 1).contiguous() else: hidden_states = input_ids # hidden_states shape change: # [batch_size, sub_seq_len, hidden_size] -> [sub_seq_len, batch_size, hidden_size] for idx, layer in enumerate(self.bert_layers): hidden_states = layer(hidden_states, attention_masks) if self.last_stage: hidden_states = hidden_states.transpose(0, 1).contiguous() output = self.layer_norm(hidden_states) output = self.head(output, self.word_embeddings.weight, lm_labels) else: output = hidden_states # hidden_states: [sub_seq_len, batch_size, hidden_size] # word_embedding: [vocab_size, hidden_size] return output def _filter_kwargs(func, kwargs): sig = inspect.signature(func) return {k: v for k, v in kwargs.items() if k in sig.parameters} def build_pipeline_bert(num_layers, num_chunks, device=torch.device("cuda"), **kwargs): logger = get_dist_logger() pipeline_size = gpc.get_world_size(ParallelMode.PIPELINE) pipeline_rank = gpc.get_local_rank(ParallelMode.PIPELINE) rank = gpc.get_global_rank() wrapper = PipelineSharedModuleWrapper([0, pipeline_size - 1]) parts = partition_uniform(num_layers, pipeline_size, num_chunks)[pipeline_rank] models = [] for start, end in parts: kwargs["num_layers"] = num_layers kwargs["start_idx"] = start kwargs["end_idx"] = end kwargs["first_stage"] = start == 0 kwargs["last_stage"] = end == num_layers logger.info(f"Rank{rank} build layer {start}-{end}, {end-start}/{num_layers} layers") chunk = PipelineBertForPretrain(**_filter_kwargs(PipelineBertForPretrain.__init__, kwargs)).to(device) if start == 0: wrapper.register_module(chunk.embedding.word_embeddings) elif end == num_layers: wrapper.register_module(chunk.word_embeddings) models.append(chunk) if len(models) == 1: model = models[0] else: model = nn.ModuleList(models) return model
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/__init__.py
examples/tutorial/sequence_parallel/model/__init__.py
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/pooler.py
examples/tutorial/sequence_parallel/model/layers/pooler.py
import torch import torch.nn as nn from .linear import Linear class Pooler(nn.Module): """Pooler layer. Pool hidden states of a specific token (for example start of the sequence) and add a linear transformation followed by a tanh. Arguments: hidden_size: hidden size init_method: weight initialization method for the linear layer. bias is set to zero. """ def __init__(self, hidden_size): super(Pooler, self).__init__() self.dense = Linear(hidden_size, hidden_size) def forward(self, hidden_states, sequence_index=0): # hidden_states: [b, s, h] # sequence_index: index of the token to pool. pooled = hidden_states[:, sequence_index, :] pooled = self.dense(pooled) pooled = torch.tanh(pooled) return pooled
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/init_method.py
examples/tutorial/sequence_parallel/model/layers/init_method.py
import math import torch def init_normal(tensor, sigma): """Init method based on N(0, sigma).""" torch.nn.init.normal_(tensor, mean=0.0, std=sigma) def output_init_normal(tensor, sigma, num_layers): """Init method based on N(0, sigma/sqrt(2*num_layers).""" std = sigma / math.sqrt(2.0 * num_layers) torch.nn.init.normal_(tensor, mean=0.0, std=std)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/bert_layer.py
examples/tutorial/sequence_parallel/model/layers/bert_layer.py
import torch import torch.nn as nn from colossalai.kernel.jit import bias_dropout_add_fused_inference, bias_dropout_add_fused_train from colossalai.legacy.nn.layer.parallel_sequence import TransformerSelfAttentionRing from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from .dropout import get_bias_dropout_add from .mlp import TransformerMLP def attention_mask_func(attention_scores, attention_mask): attention_scores.masked_fill_(attention_mask, -10000.0) return attention_scores class BertLayer(nn.Module): """A single transformer layer. Transformer layer takes input with size [b, s, h] and returns an output of the same size. """ def __init__( self, layer_number, hidden_size, num_attention_heads, attention_dropout, mlp_ratio, hidden_dropout, is_naive_fp16, apply_residual_connection_post_layernorm=False, fp32_residual_connection=False, bias_dropout_fusion: bool = True, convert_fp16_to_fp32_in_softmax: bool = False, ): super().__init__() self.layer_number = layer_number self.apply_residual_connection_post_layernorm = apply_residual_connection_post_layernorm self.fp32_residual_connection = fp32_residual_connection # Layernorm on the input data. self.input_layernorm = LayerNorm(hidden_size) # Self attention. self.self_attention = TransformerSelfAttentionRing( hidden_size=hidden_size, num_attention_heads=num_attention_heads, attention_dropout=attention_dropout, attention_mask_func=attention_mask_func, layer_number=layer_number, apply_query_key_layer_scaling=True, convert_fp16_to_fp32_in_softmax=convert_fp16_to_fp32_in_softmax, fp16=is_naive_fp16, ) self.hidden_dropout = hidden_dropout self.bias_dropout_fusion = bias_dropout_fusion # Layernorm on the attention output self.post_attention_layernorm = LayerNorm(hidden_size) self.mlp = TransformerMLP(hidden_size=hidden_size, mlp_ratio=mlp_ratio) def forward(self, hidden_states, attention_mask): # hidden_states: [batch_size, sub_seq_len, hidden_size] # attention_mask: [batch_size, 1, sub_seq_len, seq_len] # Layer norm at the beginning of the transformer layer. layernorm_output = self.input_layernorm(hidden_states) # Self attention. attention_output, attention_bias = self.self_attention(layernorm_output, attention_mask) # Residual connection. if self.apply_residual_connection_post_layernorm: residual = layernorm_output else: residual = hidden_states # jit scripting for a nn.module (with dropout) is not # trigerring the fusion kernel. For now, we use two # different nn.functional routines to account for varying # dropout semantics during training and inference phases. if self.bias_dropout_fusion: if self.training: bias_dropout_add_func = bias_dropout_add_fused_train else: bias_dropout_add_func = bias_dropout_add_fused_inference else: bias_dropout_add_func = get_bias_dropout_add(self.training) # re-enable torch grad to enable fused optimization. with torch.enable_grad(): layernorm_input = bias_dropout_add_func( attention_output, attention_bias.expand_as(residual), residual, self.hidden_dropout ) # Layer norm post the self attention. layernorm_output = self.post_attention_layernorm(layernorm_input) # MLP. mlp_output, mlp_bias = self.mlp(layernorm_output) # Second residual connection. if self.apply_residual_connection_post_layernorm: residual = layernorm_output else: residual = layernorm_input # re-enable torch grad to enable fused optimization. with torch.enable_grad(): output = bias_dropout_add_func(mlp_output, mlp_bias.expand_as(residual), residual, self.hidden_dropout) return output
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/mlp.py
examples/tutorial/sequence_parallel/model/layers/mlp.py
import torch.nn as nn import torch.nn.functional as F from colossalai.kernel.jit import bias_gelu_impl from .linear import Linear class TransformerMLP(nn.Module): """MLP. MLP will take the input with h hidden state, project it to 4*h hidden dimension, perform nonlinear transformation, and project the state back into h hidden dimension. At the end, dropout is also applied. """ def __init__(self, hidden_size, mlp_ratio, fuse_gelu=True): super(TransformerMLP, self).__init__() # Project to 4h. self.dense_h_to_4h = Linear(hidden_size, int(hidden_size * mlp_ratio), skip_bias_add=True) self.bias_gelu_fusion = fuse_gelu self.activation_func = F.gelu # Project back to h. self.dense_4h_to_h = Linear(int(hidden_size * mlp_ratio), hidden_size, skip_bias_add=True) def forward(self, hidden_states): # hidden states should be in the shape of [s, b, h] # it will be projects into [s, b, 4h] # and projected back to [s, b, h] intermediate_parallel, bias_parallel = self.dense_h_to_4h(hidden_states) if self.bias_gelu_fusion: intermediate_parallel = bias_gelu_impl(intermediate_parallel, bias_parallel) else: intermediate_parallel = self.activation_func(intermediate_parallel + bias_parallel) # [s, b, h] output, output_bias = self.dense_4h_to_h(intermediate_parallel) return output, output_bias
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/dropout.py
examples/tutorial/sequence_parallel/model/layers/dropout.py
import torch def bias_dropout_add(x, bias, residual, prob, training): # type: (Tensor, Tensor, Tensor, float, bool) -> Tensor out = torch.nn.functional.dropout(x + bias, p=prob, training=training) out = residual + out return out def get_bias_dropout_add(training): def _bias_dropout_add(x, bias, residual, prob): return bias_dropout_add(x, bias, residual, prob, training) return _bias_dropout_add
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/__init__.py
examples/tutorial/sequence_parallel/model/layers/__init__.py
from .bert_layer import BertLayer from .embedding import Embedding, VocabEmbedding from .head import BertDualHead from .preprocess import PreProcessor
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/embedding.py
examples/tutorial/sequence_parallel/model/layers/embedding.py
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init class VocabEmbedding(torch.nn.Module): def __init__(self, num_embeddings, embedding_dim): super(VocabEmbedding, self).__init__() # Keep the input dimensions. self.num_embeddings = num_embeddings self.embedding_dim = embedding_dim self.padding_idx = None self.max_norm = None self.norm_type = 2.0 self.scale_grad_by_freq = False self.sparse = False self._weight = None # Allocate weights and initialize. self.weight = nn.Parameter(torch.empty(self.num_embeddings, self.embedding_dim)) init.xavier_uniform_(self.weight) def forward(self, hidden_state): output = F.embedding( hidden_state, self.weight, self.padding_idx, self.max_norm, self.norm_type, self.scale_grad_by_freq, self.sparse, ) return output def __repr__(self): return f"VocabEmbedding(num_embeddings={self.num_embeddings}, " f"embedding_dim={self.embedding_dim})" class Embedding(nn.Module): """Language model embeddings. Arguments: hidden_size: hidden size vocab_size: vocabulary size max_sequence_length: maximum size of sequence. This is used for positional embedding embedding_dropout_prob: dropout probability for embeddings init_method: weight initialization method num_tokentypes: size of the token-type embeddings. 0 value will ignore this embedding """ def __init__(self, hidden_size, vocab_size, max_sequence_length, embedding_dropout_prob, num_tokentypes): super(Embedding, self).__init__() self.hidden_size = hidden_size self.num_tokentypes = num_tokentypes self.word_embeddings = VocabEmbedding(vocab_size, self.hidden_size) # Position embedding (serial). self.position_embeddings = torch.nn.Embedding(max_sequence_length, self.hidden_size) # Token type embedding. # Add this as an optional field that can be added through # method call so we can load a pretrain model without # token types and add them as needed. if self.num_tokentypes > 0: self.tokentype_embeddings = torch.nn.Embedding(self.num_tokentypes, self.hidden_size) else: self.tokentype_embeddings = None # Embeddings dropout self.embedding_dropout = torch.nn.Dropout(embedding_dropout_prob) @property def word_embedding_weight(self): return self.word_embeddings.weight def forward(self, input_ids, position_ids, tokentype_ids=None): # Embeddings. words_embeddings = self.word_embeddings(input_ids) position_embeddings = self.position_embeddings(position_ids) embeddings = words_embeddings + position_embeddings if tokentype_ids is not None and self.tokentype_embeddings is not None: embeddings = embeddings + self.tokentype_embeddings(tokentype_ids) # Dropout. embeddings = self.embedding_dropout(embeddings) return embeddings
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/head.py
examples/tutorial/sequence_parallel/model/layers/head.py
import torch import torch.nn as nn import torch.nn.functional as F from loss_func.cross_entropy import vocab_cross_entropy from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.nn.layer.layernorm import MixedFusedLayerNorm as LayerNorm from .linear import Linear from .pooler import Pooler class BertLMHead(nn.Module): """Masked LM head for Bert Arguments: hidden_size: hidden size init_method: init method for weight initialization layernorm_epsilon: tolerance for layer norm divisions """ def __init__( self, vocab_size, hidden_size, ): super(BertLMHead, self).__init__() self.bias = torch.nn.Parameter(torch.zeros(vocab_size)) self.dense = Linear(hidden_size, hidden_size) self.layernorm = LayerNorm(hidden_size) self.gelu = torch.nn.functional.gelu def forward(self, hidden_states, word_embeddings_weight, lm_labels): hidden_states = self.dense(hidden_states) hidden_states = self.gelu(hidden_states) hidden_states = self.layernorm(hidden_states) output = F.linear(hidden_states, word_embeddings_weight, self.bias) lm_loss = vocab_cross_entropy(output, lm_labels) return lm_loss class BertBinaryHead(nn.Module): def __init__(self, hidden_size): super().__init__() self.pooler = Pooler(hidden_size) self.dense = Linear(hidden_size, 2) def forward(self, hidden_states): if gpc.get_local_rank(ParallelMode.SEQUENCE) == 0: output = self.pooler(hidden_states) output = self.dense(output) else: output = None return output class BertDualHead(nn.Module): def __init__(self, hidden_size, vocab_size, add_binary_head): super().__init__() self.lm_head = BertLMHead(vocab_size, hidden_size) self.add_binary_head = add_binary_head if add_binary_head: self.binary_head = BertBinaryHead(hidden_size) else: self.binary_head = None def forward(self, hidden_states, word_embeddings_weight, lm_labels): if self.add_binary_head: binary_output = self.binary_head(hidden_states) else: binary_output = None lm_loss = self.lm_head(hidden_states, word_embeddings_weight, lm_labels) return lm_loss, binary_output
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/preprocess.py
examples/tutorial/sequence_parallel/model/layers/preprocess.py
import torch import torch.nn as nn from colossalai.legacy.context.parallel_mode import ParallelMode from colossalai.legacy.core import global_context as gpc class PreProcessor(nn.Module): def __init__(self, sub_seq_length): super().__init__() self.sub_seq_length = sub_seq_length def bert_position_ids(self, token_ids): # Create position ids seq_length = token_ids.size(1) local_rank = gpc.get_local_rank(ParallelMode.SEQUENCE) position_ids = torch.arange( seq_length * local_rank, seq_length * (local_rank + 1), dtype=torch.long, device=token_ids.device ) position_ids = position_ids.unsqueeze(0).expand_as(token_ids) return position_ids def bert_extended_attention_mask(self, attention_mask): local_rank = gpc.get_local_rank(ParallelMode.SEQUENCE) start_index = local_rank * self.sub_seq_length end_index = (local_rank + 1) * self.sub_seq_length # We create a 3D attention mask from a 2D tensor mask. # [b, 1, s] attention_mask_b1s = attention_mask.unsqueeze(1) # [b, s, 1] attention_mask_bs1 = attention_mask.unsqueeze(2) # [b, s/D, s] attention_mask_bss = attention_mask_b1s * attention_mask_bs1 attention_mask_bss = attention_mask_bss[:, start_index:end_index, :] # [b, 1, s/D, s] extended_attention_mask = attention_mask_bss.unsqueeze(1) # Convert attention mask to binary: extended_attention_mask = extended_attention_mask < 0.5 return extended_attention_mask def forward(self, input_ids=None, attention_mask=None): if attention_mask is not None: extended_attention_mask = self.bert_extended_attention_mask(attention_mask) else: extended_attention_mask = None if input_ids is not None: position_ids = self.bert_position_ids(input_ids) else: position_ids = None return position_ids, extended_attention_mask
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/model/layers/linear.py
examples/tutorial/sequence_parallel/model/layers/linear.py
import torch import torch.nn as nn import torch.nn.functional as F import torch.nn.init as init from torch.nn import Parameter class Linear(nn.Module): """Linear layer with column parallelism. The linear layer is defined as Y = XA + b. A is parallelized along its second dimension as A = [A_1, ..., A_p]. Arguments: input_size: first dimension of matrix A. output_size: second dimension of matrix A. bias: If true, add bias init_method: method to initialize weights. Note that bias is always set to zero. stride: For the strided linear layers. keep_master_weight_for_test: This was added for testing and should be set to False. It returns the master weights used for initialization. skip_bias_add: This was added to enable performance optimations where bias can be fused with other elementwise operations. we skip adding bias but instead return it. """ def __init__(self, input_size, output_size, bias=True, skip_bias_add=False): super(Linear, self).__init__() # Keep input parameters self.input_size = input_size self.output_size = output_size self.skip_bias_add = skip_bias_add self.weight = Parameter( torch.empty( self.output_size, self.input_size, ) ) init.normal_(self.weight) if bias: self.bias = Parameter(torch.empty(self.output_size)) # Always initialize bias to zero. with torch.no_grad(): self.bias.zero_() else: self.register_parameter("bias", None) def forward(self, input_): # Matrix multiply. bias = self.bias if not self.skip_bias_add else None output = F.linear(input_, self.weight, bias) if self.skip_bias_add: return output, self.bias else: return output def __repr__(self): return ( f"Linear(in_features={self.input_size}, out_features={self.output_size}, " + f"bias={self.bias is not None}, skip_bias_add={self.skip_bias_add})" )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/lr_scheduler/annealing_lr.py
examples/tutorial/sequence_parallel/lr_scheduler/annealing_lr.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Learning rate decay functions.""" import math class AnnealingLR(object): """Anneals the learning rate.""" def __init__( self, optimizer, max_lr, min_lr, warmup_steps, decay_steps, decay_style, use_checkpoint_lr_scheduler=True, override_lr_scheduler=False, ): # Class values. self.optimizer = optimizer self.max_lr = float(max_lr) self.min_lr = min_lr assert self.min_lr >= 0.0 assert self.max_lr >= self.min_lr self.warmup_steps = warmup_steps self.num_steps = 0 self.decay_steps = decay_steps assert self.decay_steps > 0 assert self.warmup_steps < self.decay_steps self.decay_style = decay_style self.override_lr_scheduler = override_lr_scheduler self.use_checkpoint_lr_scheduler = use_checkpoint_lr_scheduler if self.override_lr_scheduler: assert not self.use_checkpoint_lr_scheduler, "both override and " "use-checkpoint are set." # Set the learning rate self.step(0) def get_lr(self): """Learning rate decay functions from: https://openreview.net/pdf?id=BJYwwY9ll pg. 4""" # Use linear warmup for the initial part. if self.warmup_steps > 0 and self.num_steps <= self.warmup_steps: return self.max_lr * float(self.num_steps) / float(self.warmup_steps) # If the learning rate is constant, just return the initial value. if self.decay_style == "constant": return self.max_lr # For any steps larger than `self.decay_steps`, use `self.min_lr`. if self.num_steps > self.decay_steps: return self.min_lr # If we are done with the warmup period, use the decay style. num_steps_ = self.num_steps - self.warmup_steps decay_steps_ = self.decay_steps - self.warmup_steps decay_ratio = float(num_steps_) / float(decay_steps_) assert decay_ratio >= 0.0 assert decay_ratio <= 1.0 delta_lr = self.max_lr - self.min_lr if self.decay_style == "linear": coeff = 1.0 - decay_ratio elif self.decay_style == "cosine": coeff = 0.5 * (math.cos(math.pi * decay_ratio) + 1.0) else: raise Exception("{} decay style is not supported.".format(self.decay_style)) return self.min_lr + coeff * delta_lr def step(self, increment=1): """Set lr for all parameters groups.""" self.num_steps += increment new_lr = self.get_lr() for group in self.optimizer.param_groups: group["lr"] = new_lr def state_dict(self): state_dict = { "max_lr": self.max_lr, "warmup_steps": self.warmup_steps, "num_steps": self.num_steps, "decay_style": self.decay_style, "decay_steps": self.decay_steps, "min_lr": self.min_lr, } return state_dict def _check_and_set(self, cls_value, sd_value, name): """Auxiliary function for checking the values in the checkpoint and setting them.""" if self.override_lr_scheduler: return cls_value if not self.use_checkpoint_lr_scheduler: assert cls_value == sd_value, ( f"AnnealingLR: class input value {cls_value} and checkpoint" f"value {sd_value} for {name} do not match" ) return sd_value def load_state_dict(self, sd): if "start_lr" in sd: max_lr_ = sd["start_lr"] else: max_lr_ = sd["max_lr"] self.max_lr = self._check_and_set(self.max_lr, max_lr_, "learning rate") self.min_lr = self._check_and_set(self.min_lr, sd["min_lr"], "minimum learning rate") if "warmup_iter" in sd: warmup_steps_ = sd["warmup_iter"] else: warmup_steps_ = sd["warmup_steps"] self.warmup_steps = self._check_and_set(self.warmup_steps, warmup_steps_, "warmup iterations") if "end_iter" in sd: decay_steps_ = sd["end_iter"] else: decay_steps_ = sd["decay_steps"] self.decay_steps = self._check_and_set(self.decay_steps, decay_steps_, "total number of iterations") self.decay_style = self._check_and_set(self.decay_style, sd["decay_style"], "decay style") if "num_iters" in sd: num_steps = sd["num_iters"] else: num_steps = sd["num_steps"] self.step(increment=num_steps)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/lr_scheduler/__init__.py
examples/tutorial/sequence_parallel/lr_scheduler/__init__.py
from .annealing_lr import AnnealingLR
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/bert_helper.py
examples/tutorial/sequence_parallel/data/bert_helper.py
import torch from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc _MAX_DATA_DIM = 5 def _build_key_size_numel_dictionaries(keys, data): """Build the size on rank 0 and broadcast.""" max_dim = _MAX_DATA_DIM sizes = [0 for _ in range(max_dim) for _ in keys] # Pack the sizes on rank zero. if not gpc.is_initialized(ParallelMode.TENSOR) or gpc.get_local_rank(ParallelMode.TENSOR) == 0: offset = 0 for key in keys: assert data[key].dim() < max_dim, "you should increase MAX_DATA_DIM" size = data[key].size() for i, s in enumerate(size): sizes[i + offset] = s offset += max_dim # Move to GPU and broadcast. sizes_cuda = torch.cuda.LongTensor(sizes) torch.distributed.broadcast( sizes_cuda, gpc.get_ranks_in_group(ParallelMode.TENSOR)[0], group=gpc.get_group(ParallelMode.TENSOR) ) # Move back to cpu and unpack. sizes_cpu = sizes_cuda.cpu() key_size = {} key_numel = {} total_numel = 0 offset = 0 for key in keys: i = 0 size = [] numel = 1 while sizes_cpu[offset + i] > 0: this_size = sizes_cpu[offset + i] size.append(this_size) numel *= this_size i += 1 key_size[key] = size key_numel[key] = numel total_numel += numel offset += max_dim return key_size, key_numel, total_numel def broadcast_data(keys, data, datatype): """Broadcast data from rank zero of each model parallel group to the members of the same model parallel group. Arguments: keys: list of keys in the data dictionary to be broadcasted data: data dictionary of string keys and cpu tensor values. datatype: torch data type of all tensors in data associated with keys. """ # Build (key, size) and (key, number of elements) dictionaries along # with the total number of elements on all ranks. key_size, key_numel, total_numel = _build_key_size_numel_dictionaries(keys, data) # Pack on rank zero. if not gpc.is_initialized(ParallelMode.TENSOR) or gpc.get_local_rank(ParallelMode.TENSOR) == 0: # Check that all keys have the same data type. # Flatten the data associated with the keys flatten_data = torch.cat([data[key].contiguous().view(-1) for key in keys], dim=0).cuda() else: flatten_data = torch.empty(total_numel, device=torch.cuda.current_device(), dtype=datatype) # Broadcast torch.distributed.broadcast( flatten_data, gpc.get_ranks_in_group(ParallelMode.TENSOR)[0], group=gpc.get_group(ParallelMode.TENSOR) ) # Unpack output = {} offset = 0 for key in keys: size = key_size[key] numel = key_numel[key] output[key] = flatten_data.narrow(0, offset, numel).view(size) offset += numel return output def get_batch(data_iterator): """Build the batch.""" # Items and their type. keys = ["text", "types", "labels", "is_random", "loss_mask", "padding_mask"] datatype = torch.int64 # Broadcast data. if data_iterator is not None: data = next(data_iterator) else: data = None data_b = broadcast_data(keys, data, datatype) # Unpack. tokens = data_b["text"].long() types = data_b["types"].long() sentence_order = data_b["is_random"].long() loss_mask = data_b["loss_mask"].float() lm_labels = data_b["labels"].long() padding_mask = data_b["padding_mask"].long() return tokens, types, sentence_order, loss_mask, lm_labels, padding_mask def get_batch_for_sequence_parallel(data_iterator): """Build the batch.""" # Items and their type. keys = ["text", "types", "labels", "is_random", "loss_mask", "padding_mask"] datatype = torch.int64 # Broadcast data. if data_iterator is not None: data = next(data_iterator) else: data = None # unpack data_b = broadcast_data(keys, data, datatype) # # get tensor parallel local rank global_rank = torch.distributed.get_rank() local_world_size = 1 if not gpc.is_initialized(ParallelMode.TENSOR) else gpc.get_world_size(ParallelMode.TENSOR) local_rank = global_rank % local_world_size seq_length = data_b["text"].size(1) sub_seq_length = seq_length // local_world_size sub_seq_start = local_rank * sub_seq_length sub_seq_end = (local_rank + 1) * sub_seq_length # # # Unpack. tokens = data_b["text"][:, sub_seq_start:sub_seq_end].long() types = data_b["types"][:, sub_seq_start:sub_seq_end].long() sentence_order = data_b["is_random"].long() loss_mask = data_b["loss_mask"][:, sub_seq_start:sub_seq_end].float() lm_labels = data_b["labels"][:, sub_seq_start:sub_seq_end].long() padding_mask = data_b["padding_mask"].long() return tokens, types, sentence_order, loss_mask, lm_labels, padding_mask class SequenceParallelDataIterator: def __init__(self, data_iter): self.data_iter = data_iter def __iter__(self): return self.data_iter def __next__(self): return get_batch_for_sequence_parallel(self.data_iter)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/__init__.py
examples/tutorial/sequence_parallel/data/__init__.py
import torch from colossalai.legacy.context import ParallelMode from colossalai.legacy.context.parallel_context import ParallelContext from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from .datasets.builder import build_train_valid_test_datasets from .datasets.data_samplers import build_pretraining_data_loader def cyclic_iter(iter): while True: for x in iter: yield x def build_train_valid_test_data_iterators( train_iters, global_batch_size, eval_interval, eval_iters, dataloader_type="single", **kwargs ): (train_dataloader, valid_dataloader, test_dataloader) = (None, None, None) logger = get_dist_logger() logger.info("> building train, validation, and test datasets ...", ranks=[0]) # Backward compatibility, assume fixed batch size. # if iteration > 0 and consumed_train_samples == 0: # assert train_samples is None, \ # 'only backward compatibility support for iteration-based training' # consumed_train_samples = iteration * global_batch_size # if iteration > 0 and consumed_valid_samples == 0: # if train_samples is None: # consumed_valid_samples = (iteration // eval_interval) * \ # eval_iters * global_batch_size # Data loader only on rank 0 of each model parallel group. if not gpc.is_initialized(ParallelMode.TENSOR) or gpc.get_local_rank(ParallelMode.TENSOR) == 0: # Number of train/valid/test samples. train_samples = train_iters * global_batch_size eval_iters_ = (train_iters // eval_interval + 1) * eval_iters test_iters = eval_iters train_val_test_num_samples = [train_samples, eval_iters_ * global_batch_size, test_iters * global_batch_size] logger.info(" > datasets target sizes (minimum size):") logger.info(" train: {}".format(train_val_test_num_samples[0]), ranks=[0]) logger.info(" validation: {}".format(train_val_test_num_samples[1]), ranks=[0]) logger.info(" test: {}".format(train_val_test_num_samples[2]), ranks=[0]) # Build the datasets. train_ds, valid_ds, test_ds = build_train_valid_test_datasets( train_valid_test_num_samples=train_val_test_num_samples, **kwargs ) # Build dataloaders. dp_size = gpc.get_world_size(ParallelMode.DATA) train_dataloader = build_pretraining_data_loader( train_ds, consumed_samples=0, micro_batch_size=global_batch_size // dp_size ) valid_dataloader = build_pretraining_data_loader( valid_ds, consumed_samples=0, micro_batch_size=global_batch_size // dp_size ) test_dataloader = build_pretraining_data_loader(test_ds, 0, micro_batch_size=global_batch_size // dp_size) # Flags to know if we need to do training/validation/testing. do_train = train_dataloader is not None and train_iters > 0 do_valid = valid_dataloader is not None and eval_iters > 0 do_test = test_dataloader is not None and eval_iters > 0 # Need to broadcast num_tokens and num_type_tokens. flags = torch.cuda.LongTensor([int(do_train), int(do_valid), int(do_test)]) else: flags = torch.cuda.LongTensor([0, 0, 0]) # Broadcast num tokens. torch.distributed.broadcast( flags, gpc.get_ranks_in_group(ParallelMode.TENSOR)[0], group=gpc.get_group(ParallelMode.TENSOR) ) # Build iterators. dl_type = dataloader_type assert dl_type in ["single", "cyclic"] if train_dataloader is not None: train_data_iterator = iter(train_dataloader) if dl_type == "single" else iter(cyclic_iter(train_dataloader)) else: train_data_iterator = None if valid_dataloader is not None: valid_data_iterator = iter(valid_dataloader) if dl_type == "single" else iter(cyclic_iter(valid_dataloader)) else: valid_data_iterator = None if test_dataloader is not None: test_data_iterator = iter(test_dataloader) if dl_type == "single" else iter(cyclic_iter(test_dataloader)) else: test_data_iterator = None return train_data_iterator, valid_data_iterator, test_data_iterator
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/dummy_dataloader.py
examples/tutorial/sequence_parallel/data/dummy_dataloader.py
import torch class DummyDataloader: def __init__(self, batch_size, vocab_size, seq_length): self.batch_size = batch_size self.vocab_size = vocab_size self.seq_length = seq_length self.step = 0 def generate(self): tokens = torch.randint( low=0, high=self.vocab_size, size=( self.batch_size, self.seq_length, ), ) types = torch.randint( low=0, high=3, size=( self.batch_size, self.seq_length, ), ) sentence_order = torch.randint(low=0, high=2, size=(self.batch_size,)) loss_mask = torch.randint( low=0, high=2, size=( self.batch_size, self.seq_length, ), ) lm_labels = torch.randint(low=0, high=self.vocab_size, size=(self.batch_size, self.seq_length)) padding_mask = torch.randint(low=0, high=2, size=(self.batch_size, self.seq_length)) return dict( text=tokens, types=types, is_random=sentence_order, loss_mask=loss_mask, labels=lm_labels, padding_mask=padding_mask, ) def __iter__(self): return self def __next__(self): return self.generate()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/bert_tokenization.py
examples/tutorial/sequence_parallel/data/tokenizer/bert_tokenization.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tokenization classes.""" from __future__ import absolute_import, division, print_function import collections import re import unicodedata import six def validate_case_matches_checkpoint(do_lower_case, init_checkpoint): """Checks whether the casing config is consistent with the checkpoint name.""" # The casing has to be passed in by the user and there is no explicit check # as to whether it matches the checkpoint. The casing information probably # should have been stored in the bert_config.json file, but it's not, so # we have to heuristically detect it to validate. if not init_checkpoint: return m = re.match("^.*?([A-Za-z0-9_-]+)/bert_model.ckpt", init_checkpoint) if m is None: return model_name = m.group(1) lower_models = [ "uncased_L-24_H-1024_A-16", "uncased_L-12_H-768_A-12", "multilingual_L-12_H-768_A-12", "chinese_L-12_H-768_A-12", ] cased_models = ["cased_L-12_H-768_A-12", "cased_L-24_H-1024_A-16", "multi_cased_L-12_H-768_A-12"] is_bad_config = False if model_name in lower_models and not do_lower_case: is_bad_config = True actual_flag = "False" case_name = "lowercased" opposite_flag = "True" if model_name in cased_models and do_lower_case: is_bad_config = True actual_flag = "True" case_name = "cased" opposite_flag = "False" if is_bad_config: raise ValueError( "You passed in `--do_lower_case=%s` with `--init_checkpoint=%s`. " "However, `%s` seems to be a %s model, so you " "should pass in `--do_lower_case=%s` so that the fine-tuning matches " "how the model was pre-training. If this error is wrong, please " "just comment out this check." % (actual_flag, init_checkpoint, model_name, case_name, opposite_flag) ) def convert_to_unicode(text): """Converts `text` to Unicode (if it's not already), assuming utf-8 input.""" if six.PY3: if isinstance(text, str): return text elif isinstance(text, bytes): return text.decode("utf-8", "ignore") else: raise ValueError("Unsupported string type: %s" % (type(text))) elif six.PY2: if isinstance(text, str): return text.decode("utf-8", "ignore") elif isinstance(text, unicode): return text else: raise ValueError("Unsupported string type: %s" % (type(text))) else: raise ValueError("Not running on Python2 or Python 3?") def printable_text(text): """Returns text encoded in a way suitable for print or `tf.logging`.""" # These functions want `str` for both Python2 and Python3, but in one case # it's a Unicode string and in the other it's a byte string. if six.PY3: if isinstance(text, str): return text elif isinstance(text, bytes): return text.decode("utf-8", "ignore") else: raise ValueError("Unsupported string type: %s" % (type(text))) elif six.PY2: if isinstance(text, str): return text elif isinstance(text, unicode): return text.encode("utf-8") else: raise ValueError("Unsupported string type: %s" % (type(text))) else: raise ValueError("Not running on Python2 or Python 3?") def load_vocab(vocab_file): """Loads a vocabulary file into a dictionary.""" vocab = collections.OrderedDict() index = 0 with open(vocab_file, "r") as reader: while True: token = convert_to_unicode(reader.readline()) if not token: break token = token.strip() vocab[token] = index index += 1 return vocab def convert_by_vocab(vocab, items): """Converts a sequence of [tokens|ids] using the vocab.""" output = [] for item in items: output.append(vocab[item]) return output def convert_tokens_to_ids(vocab, tokens): return convert_by_vocab(vocab, tokens) def convert_ids_to_tokens(inv_vocab, ids): return convert_by_vocab(inv_vocab, ids) def whitespace_tokenize(text): """Runs basic whitespace cleaning and splitting on a piece of text.""" text = text.strip() if not text: return [] tokens = text.split() return tokens class FullTokenizer(object): """Runs end-to-end tokenization.""" def __init__(self, vocab_file, do_lower_case=True): self.vocab = load_vocab(vocab_file) self.inv_vocab = {v: k for k, v in self.vocab.items()} self.basic_tokenizer = BasicTokenizer(do_lower_case=do_lower_case) self.wordpiece_tokenizer = WordpieceTokenizer(vocab=self.vocab) def tokenize(self, text): split_tokens = [] for token in self.basic_tokenizer.tokenize(text): for sub_token in self.wordpiece_tokenizer.tokenize(token): split_tokens.append(sub_token) return split_tokens def convert_tokens_to_ids(self, tokens): return convert_by_vocab(self.vocab, tokens) def convert_ids_to_tokens(self, ids): return convert_by_vocab(self.inv_vocab, ids) @staticmethod def convert_tokens_to_string(tokens, clean_up_tokenization_spaces=True): """Converts a sequence of tokens (string) in a single string.""" def clean_up_tokenization(out_string): """Clean up a list of simple English tokenization artifacts like spaces before punctuations and abbreviated forms. """ out_string = ( out_string.replace(" .", ".") .replace(" ?", "?") .replace(" !", "!") .replace(" ,", ",") .replace(" ' ", "'") .replace(" n't", "n't") .replace(" 'm", "'m") .replace(" 's", "'s") .replace(" 've", "'ve") .replace(" 're", "'re") ) return out_string text = " ".join(tokens).replace(" ##", "").strip() if clean_up_tokenization_spaces: clean_text = clean_up_tokenization(text) return clean_text else: return text def vocab_size(self): return len(self.vocab) class BasicTokenizer(object): """Runs basic tokenization (punctuation splitting, lower casing, etc.).""" def __init__(self, do_lower_case=True): """Constructs a BasicTokenizer. Args: do_lower_case: Whether to lower case the input. """ self.do_lower_case = do_lower_case def tokenize(self, text): """Tokenizes a piece of text.""" text = convert_to_unicode(text) text = self._clean_text(text) # This was added on November 1st, 2018 for the multilingual and Chinese # models. This is also applied to the English models now, but it doesn't # matter since the English models were not trained on any Chinese data # and generally don't have any Chinese data in them (there are Chinese # characters in the vocabulary because Wikipedia does have some Chinese # words in the English Wikipedia.). text = self._tokenize_chinese_chars(text) orig_tokens = whitespace_tokenize(text) split_tokens = [] for token in orig_tokens: if self.do_lower_case: token = token.lower() token = self._run_strip_accents(token) split_tokens.extend(self._run_split_on_punc(token)) output_tokens = whitespace_tokenize(" ".join(split_tokens)) return output_tokens def _run_strip_accents(self, text): """Strips accents from a piece of text.""" text = unicodedata.normalize("NFD", text) output = [] for char in text: cat = unicodedata.category(char) if cat == "Mn": continue output.append(char) return "".join(output) def _run_split_on_punc(self, text): """Splits punctuation on a piece of text.""" chars = list(text) i = 0 start_new_word = True output = [] while i < len(chars): char = chars[i] if _is_punctuation(char): output.append([char]) start_new_word = True else: if start_new_word: output.append([]) start_new_word = False output[-1].append(char) i += 1 return ["".join(x) for x in output] def _tokenize_chinese_chars(self, text): """Adds whitespace around any CJK character.""" output = [] for char in text: cp = ord(char) if self._is_chinese_char(cp): output.append(" ") output.append(char) output.append(" ") else: output.append(char) return "".join(output) def _is_chinese_char(self, cp): """Checks whether CP is the codepoint of a CJK character.""" # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0x4E00 and cp <= 0x9FFF) or (cp >= 0x3400 and cp <= 0x4DBF) # or (cp >= 0x20000 and cp <= 0x2A6DF) # or (cp >= 0x2A700 and cp <= 0x2B73F) # or (cp >= 0x2B740 and cp <= 0x2B81F) # or (cp >= 0x2B820 and cp <= 0x2CEAF) # or (cp >= 0xF900 and cp <= 0xFAFF) or (cp >= 0x2F800 and cp <= 0x2FA1F) # ): # return True return False def _clean_text(self, text): """Performs invalid character removal and whitespace cleanup on text.""" output = [] for char in text: cp = ord(char) if cp == 0 or cp == 0xFFFD or _is_control(char): continue if _is_whitespace(char): output.append(" ") else: output.append(char) return "".join(output) class WordpieceTokenizer(object): """Runs WordPiece tokenization.""" def __init__(self, vocab, unk_token="[UNK]", max_input_chars_per_word=200): self.vocab = vocab self.unk_token = unk_token self.max_input_chars_per_word = max_input_chars_per_word def tokenize(self, text): """Tokenizes a piece of text into its word pieces. This uses a greedy longest-match-first algorithm to perform tokenization using the given vocabulary. For example: input = "unaffable" output = ["un", "##aff", "##able"] Args: text: A single token or whitespace separated tokens. This should have already been passed through `BasicTokenizer. Returns: A list of wordpiece tokens. """ text = convert_to_unicode(text) output_tokens = [] for token in whitespace_tokenize(text): chars = list(token) if len(chars) > self.max_input_chars_per_word: output_tokens.append(self.unk_token) continue is_bad = False start = 0 sub_tokens = [] while start < len(chars): end = len(chars) cur_substr = None while start < end: substr = "".join(chars[start:end]) if start > 0: substr = "##" + substr if substr in self.vocab: cur_substr = substr break end -= 1 if cur_substr is None: is_bad = True break sub_tokens.append(cur_substr) start = end if is_bad: output_tokens.append(self.unk_token) else: output_tokens.extend(sub_tokens) return output_tokens def _is_whitespace(char): """Checks whether `chars` is a whitespace character.""" # \t, \n, and \r are technically control characters but we treat them # as whitespace since they are generally considered as such. if char == " " or char == "\t" or char == "\n" or char == "\r": return True cat = unicodedata.category(char) if cat == "Zs": return True return False def _is_control(char): """Checks whether `chars` is a control character.""" # These are technically control characters but we count them as whitespace # characters. if char == "\t" or char == "\n" or char == "\r": return False cat = unicodedata.category(char) if cat in ("Cc", "Cf"): return True return False def _is_punctuation(char): """Checks whether `chars` is a punctuation character.""" cp = ord(char) # We treat all non-letter/number ASCII as punctuation. # Characters such as "^", "$", and "`" are not in the Unicode # Punctuation class but we treat them as punctuation anyways, for # consistency. if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True cat = unicodedata.category(char) if cat.startswith("P"): return True return False
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/__init__.py
examples/tutorial/sequence_parallel/data/tokenizer/__init__.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .tokenizer import build_tokenizer _TOKENIZER = None _PADDED_VOCAB_SIZE = -1 def initialize_tokenizer(vocab_file, tokenizer_type, vocab_extra_ids=0): tokenizer, padded_vocab_size = build_tokenizer(vocab_file, tokenizer_type, vocab_extra_ids) global _TOKENIZER, _PADDED_VOCAB_SIZE _TOKENIZER = tokenizer _PADDED_VOCAB_SIZE = padded_vocab_size def get_tokenizer(): global _TOKENIZER return _TOKENIZER def get_padded_vocab_size(): global _PADDED_VOCAB_SIZE return _PADDED_VOCAB_SIZE
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/tokenizer/tokenizer.py
examples/tutorial/sequence_parallel/data/tokenizer/tokenizer.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Megatron tokenizers.""" from abc import ABC, abstractmethod from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from .bert_tokenization import FullTokenizer as FullBertTokenizer def build_tokenizer(vocab_file, tokenizer_type, vocab_extra_ids=0): """Initialize tokenizer.""" if not gpc.is_initialized(ParallelMode.GLOBAL) or gpc.get_global_rank() == 0: print("> building {} tokenizer ...".format(tokenizer_type), flush=True) # Select and instantiate the tokenizer. if tokenizer_type == "BertWordPieceLowerCase": tokenizer = _BertWordPieceTokenizer(vocab_file=vocab_file, lower_case=True, vocab_extra_ids=vocab_extra_ids) elif tokenizer_type == "BertWordPieceCase": tokenizer = _BertWordPieceTokenizer(vocab_file=vocab_file, lower_case=False, vocab_extra_ids=vocab_extra_ids) else: raise NotImplementedError("{} tokenizer is not " "implemented.".format(tokenizer_type)) # Add vocab size. padded_vocab_size = _vocab_size_with_padding(tokenizer.vocab_size) return tokenizer, padded_vocab_size def _vocab_size_with_padding(orig_vocab_size, make_vocab_size_divisible_by=128): """Pad vocab size so it is divisible by model parallel size and still having GPU friendly size.""" after = orig_vocab_size if gpc.is_initialized(ParallelMode.TENSOR): multiple = make_vocab_size_divisible_by * gpc.get_world_size(ParallelMode.TENSOR) else: multiple = make_vocab_size_divisible_by while (after % multiple) != 0: after += 1 if not gpc.is_initialized(ParallelMode.GLOBAL) or gpc.get_global_rank() == 0: print( " > padded vocab (size: {}) with {} dummy tokens " "(new size: {})".format(orig_vocab_size, after - orig_vocab_size, after), flush=True, ) return after class AbstractTokenizer(ABC): """Abstract class for tokenizer.""" def __init__(self, name): self.name = name super().__init__() @property @abstractmethod def vocab_size(self): pass @property @abstractmethod def vocab(self): """Dictionary from vocab text token to id token.""" @property @abstractmethod def inv_vocab(self): """Dictionary from vocab id token to text token.""" @abstractmethod def tokenize(self, text): pass def detokenize(self, token_ids): raise NotImplementedError("detokenizer is not implemented for {} " "tokenizer".format(self.name)) @property def cls(self): raise NotImplementedError("CLS is not provided for {} " "tokenizer".format(self.name)) @property def sep(self): raise NotImplementedError("SEP is not provided for {} " "tokenizer".format(self.name)) @property def pad(self): raise NotImplementedError("PAD is not provided for {} " "tokenizer".format(self.name)) @property def eod(self): raise NotImplementedError("EOD is not provided for {} " "tokenizer".format(self.name)) @property def mask(self): raise NotImplementedError("MASK is not provided for {} " "tokenizer".format(self.name)) class _BertWordPieceTokenizer(AbstractTokenizer): """Original BERT wordpiece tokenizer.""" def __init__(self, vocab_file, lower_case=True, vocab_extra_ids=0): if lower_case: name = "BERT Lower Case" else: name = "BERT Upper Case" super().__init__(name) self.tokenizer = FullBertTokenizer(vocab_file, do_lower_case=lower_case) self.cls_id = self.tokenizer.vocab["[CLS]"] self.sep_id = self.tokenizer.vocab["[SEP]"] self.pad_id = self.tokenizer.vocab["[PAD]"] self.mask_id = self.tokenizer.vocab["[MASK]"] self._additional_special_tokens = [] # (dsachan) Add BOS and EOS tokens SPECIAL_TOKENS = {"eos_token": "[EOS]", "bos_token": "[BOS]"} self._bos_token = "[BOS]" self.add_token(self._bos_token) self._bos_token_id = self.vocab.get(self._bos_token) self._eos_token = "[EOS]" self.add_token(self._eos_token) self._eos_token_id = self.vocab.get(self._eos_token) # (dsachan) Add additional special tokens # These can be used as sentinel tokens in T5 model inputs additional_special_tokens = [] additional_special_tokens.extend(["<extra_id_{}>".format(i) for i in range(vocab_extra_ids)]) self.add_additional_special_tokens(additional_special_tokens) def add_token(self, token): if token not in self.vocab: self.inv_vocab[self.vocab_size] = token # self.vocab_size comes from len(vocab) # and it will increase as we add elements self.vocab[token] = self.vocab_size def add_additional_special_tokens(self, tokens_list): setattr(self, "additional_special_tokens", tokens_list) for value in tokens_list: self.add_token(value) @property def vocab_size(self): return self.tokenizer.vocab_size() @property def vocab(self): return self.tokenizer.vocab @property def inv_vocab(self): return self.tokenizer.inv_vocab def tokenize(self, text): text_tokens = self.tokenizer.tokenize(text) return self.tokenizer.convert_tokens_to_ids(text_tokens) def decode(self, ids): tokens = self.tokenizer.convert_ids_to_tokens(ids) return self.tokenizer.convert_tokens_to_string(tokens) def decode_token_ids(self, token_ids): tokens = self.tokenizer.convert_ids_to_tokens(token_ids) exclude_list = ["[PAD]", "[CLS]"] non_pads = [t for t in tokens if t not in exclude_list] result = "" for s in non_pads: if s.startswith("##"): result += s[2:] else: result += " " + s return result @property def cls(self): return self.cls_id @property def sep(self): return self.sep_id @property def pad(self): return self.pad_id @property def mask(self): return self.mask_id @property def bos_token(self): """Beginning of sentence token id""" return self._bos_token @property def eos_token(self): """End of sentence token id""" return self._eos_token @property def additional_special_tokens(self): """All the additional special tokens you may want to use (list of strings).""" return self._additional_special_tokens @property def bos_token_id(self): """Id of the beginning of sentence token in the vocabulary.""" return self._bos_token_id @property def eos_token_id(self): """Id of the end of sentence token in the vocabulary.""" return self._eos_token_id @property def additional_special_tokens_ids(self): """Ids of all the additional special tokens in the vocabulary (list of integers).""" return [self.vocab.get(token) for token in self._additional_special_tokens] @additional_special_tokens.setter def additional_special_tokens(self, value): self._additional_special_tokens = value
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/bert_dataset.py
examples/tutorial/sequence_parallel/data/datasets/bert_dataset.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """BERT Style dataset.""" import os import time import numpy as np import torch from torch.utils.data import Dataset from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from ..tokenizer import get_tokenizer from .dataset_utils import ( create_masked_lm_predictions, create_tokens_and_tokentypes, get_a_and_b_segments, pad_and_convert_to_numpy, truncate_segments, ) try: from . import helpers except: print("helper is not built, ignore this message if you are using synthetic data.") class BertDataset(Dataset): def __init__( self, name, indexed_dataset, data_prefix, num_epochs, max_num_samples, masked_lm_prob, max_seq_length, short_seq_prob, seed, binary_head, ): # Params to store. self.name = name self.seed = seed self.masked_lm_prob = masked_lm_prob self.max_seq_length = max_seq_length self.binary_head = binary_head # Dataset. self.indexed_dataset = indexed_dataset # Build the samples mapping. self.samples_mapping = get_samples_mapping_( self.indexed_dataset, data_prefix, num_epochs, max_num_samples, self.max_seq_length - 3, # account for added tokens, short_seq_prob, self.seed, self.name, self.binary_head, ) # Vocab stuff. tokenizer = get_tokenizer() self.vocab_id_list = list(tokenizer.inv_vocab.keys()) self.vocab_id_to_token_dict = tokenizer.inv_vocab self.cls_id = tokenizer.cls self.sep_id = tokenizer.sep self.mask_id = tokenizer.mask self.pad_id = tokenizer.pad def __len__(self): return self.samples_mapping.shape[0] def __getitem__(self, idx): start_idx, end_idx, seq_length = self.samples_mapping[idx] sample = [self.indexed_dataset[i] for i in range(start_idx, end_idx)] # Note that this rng state should be numpy and not python since # python randint is inclusive whereas the numpy one is exclusive. # We % 2**32 since numpy requires the seed to be between 0 and 2**32 - 1 np_rng = np.random.RandomState(seed=((self.seed + idx) % 2**32)) return build_training_sample( sample, seq_length, self.max_seq_length, # needed for padding self.vocab_id_list, self.vocab_id_to_token_dict, self.cls_id, self.sep_id, self.mask_id, self.pad_id, self.masked_lm_prob, np_rng, self.binary_head, ) def get_samples_mapping_( indexed_dataset, data_prefix, num_epochs, max_num_samples, max_seq_length, short_seq_prob, seed, name, binary_head ): logger = get_dist_logger() if not num_epochs: if not max_num_samples: raise ValueError("Need to specify either max_num_samples " "or num_epochs") num_epochs = np.iinfo(np.int32).max - 1 if not max_num_samples: max_num_samples = np.iinfo(np.int64).max - 1 # Filename of the index mapping indexmap_filename = data_prefix indexmap_filename += "_{}_indexmap".format(name) if num_epochs != (np.iinfo(np.int32).max - 1): indexmap_filename += "_{}ep".format(num_epochs) if max_num_samples != (np.iinfo(np.int64).max - 1): indexmap_filename += "_{}mns".format(max_num_samples) indexmap_filename += "_{}msl".format(max_seq_length) indexmap_filename += "_{:0.2f}ssp".format(short_seq_prob) indexmap_filename += "_{}s".format(seed) indexmap_filename += ".npy" # Build the indexed mapping if not exist. if torch.distributed.get_rank() == 0 and not os.path.isfile(indexmap_filename): print( " > WARNING: could not find index map file {}, building " "the indices on rank 0 ...".format(indexmap_filename) ) # Make sure the types match the helpers input types. assert indexed_dataset.doc_idx.dtype == np.int64 assert indexed_dataset.sizes.dtype == np.int32 # Build samples mapping verbose = torch.distributed.get_rank() == 0 start_time = time.time() logger.info("\n > building samples index mapping for {} ...".format(name), ranks=[0]) # First compile and then import. samples_mapping = helpers.build_mapping( indexed_dataset.doc_idx, indexed_dataset.sizes, num_epochs, max_num_samples, max_seq_length, short_seq_prob, seed, verbose, 2 if binary_head else 1, ) logger.info("\n > done building samples index maping", ranks=[0]) np.save(indexmap_filename, samples_mapping, allow_pickle=True) logger.info("\n > saved the index mapping in {}".format(indexmap_filename), ranks=[0]) # Make sure all the ranks have built the mapping logger.info( "\n > elapsed time to build and save samples mapping " "(seconds): {:4f}".format(time.time() - start_time), ranks=[0], ) # This should be a barrier but nccl barrier assumes # device_index=rank which is not the case for model # parallel case counts = torch.cuda.LongTensor([1]) torch.distributed.all_reduce(counts, group=gpc.get_group(ParallelMode.DATA)) if gpc.is_initialized(ParallelMode.PIPELINE): torch.distributed.all_reduce(counts, group=gpc.get_group(ParallelMode.PIPELINE)) assert counts[0].item() == ( torch.distributed.get_world_size() // torch.distributed.get_world_size(group=gpc.get_group(ParallelMode.SEQUENCE)) ) # Load indexed dataset. start_time = time.time() samples_mapping = np.load(indexmap_filename, allow_pickle=True, mmap_mode="r") logger.info( "\n > loading indexed mapping from {}".format(indexmap_filename) + "\n loaded indexed file in {:3.3f} seconds".format(time.time() - start_time) + "\n total number of samples: {}".format(samples_mapping.shape[0]), ranks=[0], ) return samples_mapping def build_training_sample( sample, target_seq_length, max_seq_length, vocab_id_list, vocab_id_to_token_dict, cls_id, sep_id, mask_id, pad_id, masked_lm_prob, np_rng, binary_head, ): """Build training sample. Arguments: sample: A list of sentences in which each sentence is a list token ids. target_seq_length: Desired sequence length. max_seq_length: Maximum length of the sequence. All values are padded to this length. vocab_id_list: List of vocabulary ids. Used to pick a random id. vocab_id_to_token_dict: A dictionary from vocab ids to text tokens. cls_id: Start of example id. sep_id: Separator id. mask_id: Mask token id. pad_id: Padding token id. masked_lm_prob: Probability to mask tokens. np_rng: Random number genenrator. Note that this rng state should be numpy and not python since python randint is inclusive for the opper bound whereas the numpy one is exclusive. """ if binary_head: # We assume that we have at least two sentences in the sample assert len(sample) > 1 assert target_seq_length <= max_seq_length # Divide sample into two segments (A and B). if binary_head: tokens_a, tokens_b, is_next_random = get_a_and_b_segments(sample, np_rng) else: tokens_a = [] for j in range(len(sample)): tokens_a.extend(sample[j]) tokens_b = [] is_next_random = False # Truncate to `target_sequence_length`. max_num_tokens = target_seq_length truncated = truncate_segments(tokens_a, tokens_b, len(tokens_a), len(tokens_b), max_num_tokens, np_rng) # Build tokens and toketypes. tokens, tokentypes = create_tokens_and_tokentypes(tokens_a, tokens_b, cls_id, sep_id) # Masking. max_predictions_per_seq = masked_lm_prob * max_num_tokens (tokens, masked_positions, masked_labels, _) = create_masked_lm_predictions( tokens, vocab_id_list, vocab_id_to_token_dict, masked_lm_prob, cls_id, sep_id, mask_id, max_predictions_per_seq, np_rng, ) # Padding. tokens_np, tokentypes_np, labels_np, padding_mask_np, loss_mask_np = pad_and_convert_to_numpy( tokens, tokentypes, masked_positions, masked_labels, pad_id, max_seq_length ) train_sample = { "text": tokens_np, "types": tokentypes_np, "labels": labels_np, "is_random": int(is_next_random), "loss_mask": loss_mask_np, "padding_mask": padding_mask_np, "truncated": int(truncated), } return train_sample
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/ict_dataset.py
examples/tutorial/sequence_parallel/data/datasets/ict_dataset.py
import itertools import random import numpy as np from megatron import get_args, get_tokenizer from megatron.data.dataset_utils import get_indexed_dataset_ from megatron.data.realm_dataset_utils import get_block_samples_mapping from torch.utils.data import Dataset def make_attention_mask(source_block, target_block): """ Returns a 2-dimensional (2-D) attention mask :param source_block: 1-D array :param target_block: 1-D array """ mask = (target_block[None, :] >= 1) * (source_block[:, None] >= 1) mask = mask.astype(np.int64) # (source_length, target_length) return mask def get_ict_dataset(use_titles=True, query_in_block_prob=1): """Get a dataset which uses block samples mappings to get ICT/block indexing data (via get_block()) rather than for training, since it is only built with a single epoch sample mapping. """ args = get_args() block_dataset = get_indexed_dataset_(args.data_path, "mmap", True) titles_dataset = get_indexed_dataset_(args.titles_data_path, "mmap", True) kwargs = dict( name="full", block_dataset=block_dataset, title_dataset=titles_dataset, data_prefix=args.data_path, num_epochs=1, max_num_samples=None, max_seq_length=args.seq_length, seed=1, query_in_block_prob=query_in_block_prob, use_titles=use_titles, use_one_sent_docs=args.use_one_sent_docs, ) dataset = ICTDataset(**kwargs) return dataset class ICTDataset(Dataset): """Dataset containing sentences and their blocks for an inverse cloze task.""" def __init__( self, name, block_dataset, title_dataset, data_prefix, num_epochs, max_num_samples, max_seq_length, query_in_block_prob, seed, use_titles=True, use_one_sent_docs=False, binary_head=False, ): self.name = name self.seed = seed self.max_seq_length = max_seq_length self.query_in_block_prob = query_in_block_prob self.block_dataset = block_dataset self.title_dataset = title_dataset self.rng = random.Random(self.seed) self.use_titles = use_titles self.use_one_sent_docs = use_one_sent_docs self.samples_mapping = get_block_samples_mapping( block_dataset, title_dataset, data_prefix, num_epochs, max_num_samples, max_seq_length, seed, name, use_one_sent_docs, ) self.tokenizer = get_tokenizer() self.vocab_id_list = list(self.tokenizer.inv_vocab.keys()) self.vocab_id_to_token_list = self.tokenizer.inv_vocab self.cls_id = self.tokenizer.cls self.sep_id = self.tokenizer.sep self.mask_id = self.tokenizer.mask self.pad_id = self.tokenizer.pad def __len__(self): return len(self.samples_mapping) def __getitem__(self, idx): """Get an ICT example of a pseudo-query and the block of text from which it was extracted""" sample_data = self.samples_mapping[idx] start_idx, end_idx, doc_idx, block_idx = sample_data.as_tuple() if self.use_titles: title = self.title_dataset[int(doc_idx)] title_pad_offset = 3 + len(title) else: title = None title_pad_offset = 2 block = [self.block_dataset[i] for i in range(start_idx, end_idx)] assert len(block) > 1 or self.use_one_sent_docs or self.query_in_block_prob == 1 # randint() is inclusive for Python rng rand_sent_idx = self.rng.randint(0, len(block) - 1) # keep the query in the context query_in_block_prob fraction of the time. if self.rng.random() < self.query_in_block_prob: query = block[rand_sent_idx].copy() else: query = block.pop(rand_sent_idx) # still need to truncate because blocks are concluded when # the sentence lengths have exceeded max_seq_length. query = query[: self.max_seq_length - 2] block = list(itertools.chain(*block))[: self.max_seq_length - title_pad_offset] query_tokens, query_pad_mask = self.concat_and_pad_tokens(query) context_tokens, context_pad_mask = self.concat_and_pad_tokens(block, title) query_mask = make_attention_mask(query_tokens, query_tokens) context_mask = make_attention_mask(context_tokens, context_tokens) block_data = sample_data.as_array() sample = { "query_tokens": query_tokens, "query_mask": query_mask, "query_pad_mask": query_pad_mask, "context_tokens": context_tokens, "context_mask": context_mask, "context_pad_mask": context_pad_mask, "block_data": block_data, } return sample def get_block(self, start_idx, end_idx, doc_idx): """Get the IDs for an evidence block plus the title of the corresponding document""" block = [self.block_dataset[i] for i in range(start_idx, end_idx)] title = self.title_dataset[int(doc_idx)] block = list(itertools.chain(*block))[: self.max_seq_length - (3 + len(title))] block_tokens, block_pad_mask = self.concat_and_pad_tokens(block, title) return block_tokens, block_pad_mask def get_null_block(self): """Get empty block and title - used in REALM pretraining""" block, title = [], [] block_tokens, block_pad_mask = self.concat_and_pad_tokens(block, title) return block_tokens, block_pad_mask def concat_and_pad_tokens(self, tokens, title=None): """Concat with special tokens and pad sequence to self.max_seq_length""" tokens = list(tokens) if title is None: tokens = [self.cls_id] + tokens + [self.sep_id] else: title = list(title) tokens = [self.cls_id] + title + [self.sep_id] + tokens + [self.sep_id] assert len(tokens) <= self.max_seq_length num_pad = self.max_seq_length - len(tokens) pad_mask = [1] * len(tokens) + [0] * num_pad tokens += [self.pad_id] * num_pad return np.array(tokens), np.array(pad_mask)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/data_samplers.py
examples/tutorial/sequence_parallel/data/datasets/data_samplers.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Dataloaders.""" import torch from colossalai.legacy.context import ParallelMode from colossalai.legacy.core import global_context as gpc def build_pretraining_data_loader(dataset, consumed_samples, micro_batch_size, dataloader_type="single", num_workers=0): """Build dataloader given an input dataset.""" if dataset is None: return None # Megatron sampler if dataloader_type == "single": batch_sampler = MegatronPretrainingSampler( total_samples=len(dataset), consumed_samples=consumed_samples, micro_batch_size=micro_batch_size, data_parallel_rank=gpc.get_local_rank(ParallelMode.DATA), data_parallel_size=gpc.get_world_size(ParallelMode.DATA), ) elif dataloader_type == "cyclic": batch_sampler = MegatronPretrainingRandomSampler( total_samples=len(dataset), consumed_samples=consumed_samples, micro_batch_size=micro_batch_size, data_parallel_rank=gpc.get_local_rank(ParallelMode.DATA), data_parallel_size=gpc.get_world_size(ParallelMode.DATA), ) else: raise Exception("{} dataloader type is not supported.".format(dataloader_type)) # Torch dataloader. return torch.utils.data.DataLoader(dataset, batch_sampler=batch_sampler, num_workers=num_workers, pin_memory=True) class MegatronPretrainingSampler: def __init__( self, total_samples, consumed_samples, micro_batch_size, data_parallel_rank, data_parallel_size, drop_last=True ): # Keep a copy of input params for later use. self.total_samples = total_samples self.consumed_samples = consumed_samples self.micro_batch_size = micro_batch_size self.data_parallel_rank = data_parallel_rank self.micro_batch_times_data_parallel_size = self.micro_batch_size * data_parallel_size self.drop_last = drop_last # Sanity checks. assert self.total_samples > 0, "no sample to consume: {}".format(self.total_samples) assert self.consumed_samples < self.total_samples, "no samples left to consume: {}, {}".format( self.consumed_samples, self.total_samples ) assert self.micro_batch_size > 0 assert data_parallel_size > 0 assert ( self.data_parallel_rank < data_parallel_size ), "data_parallel_rank should be smaller than data size: {}, " "{}".format( self.data_parallel_rank, data_parallel_size ) def __len__(self): return self.total_samples def get_start_end_idx(self): start_idx = self.data_parallel_rank * self.micro_batch_size end_idx = start_idx + self.micro_batch_size return start_idx, end_idx def __iter__(self): batch = [] # Last batch will be dropped if drop_last is not set False for idx in range(self.consumed_samples, self.total_samples): batch.append(idx) if len(batch) == self.micro_batch_times_data_parallel_size: start_idx, end_idx = self.get_start_end_idx() yield batch[start_idx:end_idx] batch = [] # Check the last partial batch and see drop_last is set if len(batch) > 0 and not self.drop_last: start_idx, end_idx = self.get_start_end_idx() yield batch[start_idx:end_idx] class MegatronPretrainingRandomSampler: def __init__(self, total_samples, consumed_samples, micro_batch_size, data_parallel_rank, data_parallel_size): # Keep a copy of input params for later use. self.total_samples = total_samples self.consumed_samples = consumed_samples self.micro_batch_size = micro_batch_size self.data_parallel_rank = data_parallel_rank self.data_parallel_size = data_parallel_size self.micro_batch_times_data_parallel_size = self.micro_batch_size * data_parallel_size self.last_batch_size = self.total_samples % self.micro_batch_times_data_parallel_size # Sanity checks. assert self.total_samples > 0, "no sample to consume: {}".format(self.total_samples) assert self.micro_batch_size > 0 assert data_parallel_size > 0 assert ( self.data_parallel_rank < data_parallel_size ), "data_parallel_rank should be smaller than data size: {}, " "{}".format( self.data_parallel_rank, data_parallel_size ) def __len__(self): return self.total_samples def __iter__(self): active_total_samples = self.total_samples - self.last_batch_size self.epoch = self.consumed_samples // active_total_samples current_epoch_samples = self.consumed_samples % active_total_samples assert current_epoch_samples % self.micro_batch_times_data_parallel_size == 0 # data sharding and random sampling bucket_size = (self.total_samples // self.micro_batch_times_data_parallel_size) * self.micro_batch_size bucket_offset = current_epoch_samples // self.data_parallel_size start_idx = self.data_parallel_rank * bucket_size g = torch.Generator() g.manual_seed(self.epoch) random_idx = torch.randperm(bucket_size, generator=g).tolist() idx_range = [start_idx + x for x in random_idx[bucket_offset:]] batch = [] # Last batch if not complete will be dropped. for idx in idx_range: batch.append(idx) if len(batch) == self.micro_batch_size: self.consumed_samples += self.micro_batch_times_data_parallel_size yield batch batch = []
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/dataset_utils.py
examples/tutorial/sequence_parallel/data/datasets/dataset_utils.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors, and NVIDIA. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Most of the code here has been copied from: # https://github.com/google-research/albert/blob/master/create_pretraining_data.py # with some modifications. import collections import math import time import numpy as np from colossalai.logging import get_dist_logger from .blendable_dataset import BlendableDataset from .indexed_dataset import make_dataset as make_indexed_dataset DSET_TYPE_STD = "standard_bert" DSET_TYPE_ICT = "ict" DSET_TYPES = [DSET_TYPE_ICT, DSET_TYPE_STD] def get_datasets_weights_and_num_samples(data_prefix, train_valid_test_num_samples): # The data prefix should be in the format of: # weight-1, data-prefix-1, weight-2, data-prefix-2, .. assert len(data_prefix) % 2 == 0 num_datasets = len(data_prefix) // 2 weights = [0] * num_datasets prefixes = [0] * num_datasets for i in range(num_datasets): weights[i] = float(data_prefix[2 * i]) prefixes[i] = (data_prefix[2 * i + 1]).strip() # Normalize weights weight_sum = 0.0 for weight in weights: weight_sum += weight assert weight_sum > 0.0 weights = [weight / weight_sum for weight in weights] # Add 0.5% (the 1.005 factor) so in case the bleding dataset does # not uniformly distribute the number of samples, we still have # samples left to feed to the network. datasets_train_valid_test_num_samples = [] for weight in weights: datasets_train_valid_test_num_samples.append( [int(math.ceil(val * weight * 1.005)) for val in train_valid_test_num_samples] ) return prefixes, weights, datasets_train_valid_test_num_samples def compile_helper(): """Compile helper function ar runtime. Make sure this is invoked on a single process.""" import os import subprocess path = os.path.abspath(os.path.dirname(__file__)) ret = subprocess.run(["make", "-C", path]) if ret.returncode != 0: print("Making C++ dataset helpers module failed, exiting.") import sys sys.exit(1) def get_a_and_b_segments(sample, np_rng): """Divide sample into a and b segments.""" # Number of sentences in the sample. n_sentences = len(sample) # Make sure we always have two sentences. assert n_sentences > 1, "make sure each sample has at least two sentences." # First part: # `a_end` is how many sentences go into the `A`. a_end = 1 if n_sentences >= 3: # Note that randin in numpy is exclusive. a_end = np_rng.randint(1, n_sentences) tokens_a = [] for j in range(a_end): tokens_a.extend(sample[j]) # Second part: tokens_b = [] for j in range(a_end, n_sentences): tokens_b.extend(sample[j]) # Random next: is_next_random = False if np_rng.random() < 0.5: is_next_random = True tokens_a, tokens_b = tokens_b, tokens_a return tokens_a, tokens_b, is_next_random def truncate_segments(tokens_a, tokens_b, len_a, len_b, max_num_tokens, np_rng): """Truncates a pair of sequences to a maximum sequence length.""" # print(len_a, len_b, max_num_tokens) assert len_a > 0 if len_a + len_b <= max_num_tokens: return False while len_a + len_b > max_num_tokens: if len_a > len_b: len_a -= 1 tokens = tokens_a else: len_b -= 1 tokens = tokens_b if np_rng.random() < 0.5: del tokens[0] else: tokens.pop() return True def create_tokens_and_tokentypes(tokens_a, tokens_b, cls_id, sep_id): """Merge segments A and B, add [CLS] and [SEP] and build tokentypes.""" tokens = [] tokentypes = [] # [CLS]. tokens.append(cls_id) tokentypes.append(0) # Segment A. for token in tokens_a: tokens.append(token) tokentypes.append(0) # [SEP]. tokens.append(sep_id) tokentypes.append(0) # Segment B. for token in tokens_b: tokens.append(token) tokentypes.append(1) if tokens_b: # [SEP]. tokens.append(sep_id) tokentypes.append(1) return tokens, tokentypes MaskedLmInstance = collections.namedtuple("MaskedLmInstance", ["index", "label"]) def is_start_piece(piece): """Check if the current word piece is the starting piece (BERT).""" # When a word has been split into # WordPieces, the first token does not have any marker and any subsequence # tokens are prefixed with ##. So whenever we see the ## token, we # append it to the previous set of word indexes. return not piece.startswith("##") def create_masked_lm_predictions( tokens, vocab_id_list, vocab_id_to_token_dict, masked_lm_prob, cls_id, sep_id, mask_id, max_predictions_per_seq, np_rng, max_ngrams=3, do_whole_word_mask=True, favor_longer_ngram=False, do_permutation=False, ): """Creates the predictions for the masked LM objective. Note: Tokens here are vocab ids and not text tokens.""" cand_indexes = [] # Note(mingdachen): We create a list for recording if the piece is # the starting piece of current token, where 1 means true, so that # on-the-fly whole word masking is possible. token_boundary = [0] * len(tokens) for i, token in enumerate(tokens): if token == cls_id or token == sep_id: token_boundary[i] = 1 continue # Whole Word Masking means that if we mask all of the wordpieces # corresponding to an original word. # # Note that Whole Word Masking does *not* change the training code # at all -- we still predict each WordPiece independently, softmaxed # over the entire vocabulary. if do_whole_word_mask and len(cand_indexes) >= 1 and not is_start_piece(vocab_id_to_token_dict[token]): cand_indexes[-1].append(i) else: cand_indexes.append([i]) if is_start_piece(vocab_id_to_token_dict[token]): token_boundary[i] = 1 output_tokens = list(tokens) masked_lm_positions = [] masked_lm_labels = [] if masked_lm_prob == 0: return (output_tokens, masked_lm_positions, masked_lm_labels, token_boundary) num_to_predict = min(max_predictions_per_seq, max(1, int(round(len(tokens) * masked_lm_prob)))) # Note(mingdachen): # By default, we set the probabilities to favor shorter ngram sequences. ngrams = np.arange(1, max_ngrams + 1, dtype=np.int64) pvals = 1.0 / np.arange(1, max_ngrams + 1) pvals /= pvals.sum(keepdims=True) if favor_longer_ngram: pvals = pvals[::-1] ngram_indexes = [] for idx in range(len(cand_indexes)): ngram_index = [] for n in ngrams: ngram_index.append(cand_indexes[idx : idx + n]) ngram_indexes.append(ngram_index) np_rng.shuffle(ngram_indexes) masked_lms = [] covered_indexes = set() for cand_index_set in ngram_indexes: if len(masked_lms) >= num_to_predict: break if not cand_index_set: continue # Note(mingdachen): # Skip current piece if they are covered in lm masking or previous ngrams. for index_set in cand_index_set[0]: for index in index_set: if index in covered_indexes: continue n = np_rng.choice( ngrams[: len(cand_index_set)], p=pvals[: len(cand_index_set)] / pvals[: len(cand_index_set)].sum(keepdims=True), ) index_set = sum(cand_index_set[n - 1], []) n -= 1 # Note(mingdachen): # Repeatedly looking for a candidate that does not exceed the # maximum number of predictions by trying shorter ngrams. while len(masked_lms) + len(index_set) > num_to_predict: if n == 0: break index_set = sum(cand_index_set[n - 1], []) n -= 1 # If adding a whole-word mask would exceed the maximum number of # predictions, then just skip this candidate. if len(masked_lms) + len(index_set) > num_to_predict: continue is_any_index_covered = False for index in index_set: if index in covered_indexes: is_any_index_covered = True break if is_any_index_covered: continue for index in index_set: covered_indexes.add(index) masked_token = None # 80% of the time, replace with [MASK] if np_rng.random() < 0.8: masked_token = mask_id else: # 10% of the time, keep original if np_rng.random() < 0.5: masked_token = tokens[index] # 10% of the time, replace with random word else: masked_token = vocab_id_list[np_rng.randint(0, len(vocab_id_list))] output_tokens[index] = masked_token masked_lms.append(MaskedLmInstance(index=index, label=tokens[index])) assert len(masked_lms) <= num_to_predict np_rng.shuffle(ngram_indexes) select_indexes = set() if do_permutation: for cand_index_set in ngram_indexes: if len(select_indexes) >= num_to_predict: break if not cand_index_set: continue # Note(mingdachen): # Skip current piece if they are covered in lm masking or previous ngrams. for index_set in cand_index_set[0]: for index in index_set: if index in covered_indexes or index in select_indexes: continue n = np.random.choice( ngrams[: len(cand_index_set)], p=pvals[: len(cand_index_set)] / pvals[: len(cand_index_set)].sum(keepdims=True), ) index_set = sum(cand_index_set[n - 1], []) n -= 1 while len(select_indexes) + len(index_set) > num_to_predict: if n == 0: break index_set = sum(cand_index_set[n - 1], []) n -= 1 # If adding a whole-word mask would exceed the maximum number of # predictions, then just skip this candidate. if len(select_indexes) + len(index_set) > num_to_predict: continue is_any_index_covered = False for index in index_set: if index in covered_indexes or index in select_indexes: is_any_index_covered = True break if is_any_index_covered: continue for index in index_set: select_indexes.add(index) assert len(select_indexes) <= num_to_predict select_indexes = sorted(select_indexes) permute_indexes = list(select_indexes) np_rng.shuffle(permute_indexes) orig_token = list(output_tokens) for src_i, tgt_i in zip(select_indexes, permute_indexes): output_tokens[src_i] = orig_token[tgt_i] masked_lms.append(MaskedLmInstance(index=src_i, label=orig_token[src_i])) masked_lms = sorted(masked_lms, key=lambda x: x.index) for p in masked_lms: masked_lm_positions.append(p.index) masked_lm_labels.append(p.label) return (output_tokens, masked_lm_positions, masked_lm_labels, token_boundary) def pad_and_convert_to_numpy(tokens, tokentypes, masked_positions, masked_labels, pad_id, max_seq_length): """Pad sequences and convert them to numpy.""" # Some checks. num_tokens = len(tokens) padding_length = max_seq_length - num_tokens assert padding_length >= 0 assert len(tokentypes) == num_tokens assert len(masked_positions) == len(masked_labels) # Tokens and token types. filler = [pad_id] * padding_length tokens_np = np.array(tokens + filler, dtype=np.int64) tokentypes_np = np.array(tokentypes + filler, dtype=np.int64) # Padding mask. padding_mask_np = np.array([1] * num_tokens + [0] * padding_length, dtype=np.int64) # Lables and loss mask. labels = [-1] * max_seq_length loss_mask = [0] * max_seq_length for i in range(len(masked_positions)): assert masked_positions[i] < num_tokens labels[masked_positions[i]] = masked_labels[i] loss_mask[masked_positions[i]] = 1 labels_np = np.array(labels, dtype=np.int64) loss_mask_np = np.array(loss_mask, dtype=np.int64) return tokens_np, tokentypes_np, labels_np, padding_mask_np, loss_mask_np def build_train_valid_test_datasets( data_prefix, data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type="standard_bert", ): if len(data_prefix) == 1: return _build_train_valid_test_datasets( data_prefix[0], data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type=dataset_type, ) # Blending dataset. # Parse the values. output = get_datasets_weights_and_num_samples(data_prefix, train_valid_test_num_samples) prefixes, weights, datasets_train_valid_test_num_samples = output # Build individual datasets. train_datasets = [] valid_datasets = [] test_datasets = [] for i in range(len(prefixes)): train_ds, valid_ds, test_ds = _build_train_valid_test_datasets( prefixes[i], data_impl, splits_string, datasets_train_valid_test_num_samples[i], max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type=dataset_type, ) if train_ds: train_datasets.append(train_ds) if valid_ds: valid_datasets.append(valid_ds) if test_ds: test_datasets.append(test_ds) # Blend. blending_train_dataset = None if train_datasets: blending_train_dataset = BlendableDataset(train_datasets, weights) blending_valid_dataset = None if valid_datasets: blending_valid_dataset = BlendableDataset(valid_datasets, weights) blending_test_dataset = None if test_datasets: blending_test_dataset = BlendableDataset(test_datasets, weights) return (blending_train_dataset, blending_valid_dataset, blending_test_dataset) def _build_train_valid_test_datasets( data_prefix, data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type="standard_bert", ): logger = get_dist_logger() if dataset_type not in DSET_TYPES: raise ValueError("Invalid dataset_type: ", dataset_type) # Indexed dataset. indexed_dataset = get_indexed_dataset_(data_prefix, data_impl, skip_warmup) if dataset_type == DSET_TYPE_ICT: args = get_args() title_dataset = get_indexed_dataset_(args.titles_data_path, data_impl, skip_warmup) # Get start and end indices of train/valid/train into doc-idx # Note that doc-idx is designed to be num-docs + 1 so we can # easily iterate over it. total_num_of_documents = indexed_dataset.doc_idx.shape[0] - 1 splits = get_train_valid_test_split_(splits_string, total_num_of_documents) # Print stats about the splits. logger.info("\n > dataset split:") def print_split_stats(name, index): start_index = indexed_dataset.doc_idx[splits[index]] end_index = indexed_dataset.doc_idx[splits[index + 1]] logger.info( "\n {}:".format(name) + "\n document indices in [{}, {}) total of {} documents".format( splits[index], splits[index + 1], splits[index + 1] - splits[index] ) + "\n sentence indices in [{}, {}) total of {} sentences".format( start_index, end_index, end_index - start_index ), ranks=[0], ) print_split_stats("train", 0) print_split_stats("validation", 1) print_split_stats("test", 2) def build_dataset(index, name): from .bert_dataset import BertDataset dataset = None if splits[index + 1] > splits[index]: # Get the pointer to the original doc-idx so we can set it later. doc_idx_ptr = indexed_dataset.get_doc_idx() # Slice the doc-idx start_index = splits[index] # Add +1 so we can index into the dataset to get the upper bound. end_index = splits[index + 1] + 1 # New doc_idx view. indexed_dataset.set_doc_idx(doc_idx_ptr[start_index:end_index]) # Build the dataset accordingly. kwargs = dict( name=name, data_prefix=data_prefix, num_epochs=None, max_num_samples=train_valid_test_num_samples[index], max_seq_length=max_seq_length, seed=seed, binary_head=binary_head, ) if dataset_type == DSET_TYPE_ICT: args = get_args() dataset = ICTDataset( block_dataset=indexed_dataset, title_dataset=title_dataset, query_in_block_prob=args.query_in_block_prob, use_one_sent_docs=args.use_one_sent_docs, **kwargs, ) else: dataset = BertDataset( indexed_dataset=indexed_dataset, masked_lm_prob=masked_lm_prob, short_seq_prob=short_seq_prob, **kwargs, ) # Set the original pointer so dataset remains the main dataset. indexed_dataset.set_doc_idx(doc_idx_ptr) # Checks. assert indexed_dataset.doc_idx[0] == 0 assert indexed_dataset.doc_idx.shape[0] == (total_num_of_documents + 1) return dataset train_dataset = build_dataset(0, "train") valid_dataset = build_dataset(1, "valid") test_dataset = build_dataset(2, "test") return (train_dataset, valid_dataset, test_dataset) def get_indexed_dataset_(data_prefix, data_impl, skip_warmup): logger = get_dist_logger() start_time = time.time() indexed_dataset = make_indexed_dataset(data_prefix, data_impl, skip_warmup) assert indexed_dataset.sizes.shape[0] == indexed_dataset.doc_idx[-1] logger.info("\n > building dataset index ...", ranks=[0]) logger.info( "\n > finished creating indexed dataset in {:4f} " "seconds".format(time.time() - start_time), ranks=[0] ) logger.info( "\n > indexed dataset stats:" + "\n number of documents: {}".format(indexed_dataset.doc_idx.shape[0] - 1) + "\n number of sentences: {}".format(indexed_dataset.sizes.shape[0]), ranks=[0], ) return indexed_dataset def get_train_valid_test_split_(splits_string, size): """Get dataset splits from comma or '/' separated string list.""" splits = [] if splits_string.find(",") != -1: splits = [float(s) for s in splits_string.split(",")] elif splits_string.find("/") != -1: splits = [float(s) for s in splits_string.split("/")] else: splits = [float(splits_string)] while len(splits) < 3: splits.append(0.0) splits = splits[:3] splits_sum = sum(splits) assert splits_sum > 0.0 splits = [split / splits_sum for split in splits] splits_index = [0] for index, split in enumerate(splits): splits_index.append(splits_index[index] + int(round(split * float(size)))) diff = splits_index[-1] - size for index in range(1, len(splits_index)): splits_index[index] -= diff assert len(splits_index) == 4 assert splits_index[-1] == size return splits_index
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/blendable_dataset.py
examples/tutorial/sequence_parallel/data/datasets/blendable_dataset.py
# coding=utf-8 # Copyright (c) 2020, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Blendable dataset.""" import time import numpy as np import torch class BlendableDataset(torch.utils.data.Dataset): def __init__(self, datasets, weights): self.datasets = datasets num_datasets = len(datasets) assert num_datasets == len(weights) self.size = 0 for dataset in self.datasets: self.size += len(dataset) # Normalize weights. weights = np.array(weights, dtype=np.float64) sum_weights = np.sum(weights) assert sum_weights > 0.0 weights /= sum_weights # Build indices. start_time = time.time() assert num_datasets < 255 self.dataset_index = np.zeros(self.size, dtype=np.uint8) self.dataset_sample_index = np.zeros(self.size, dtype=np.int64) from . import helpers helpers.build_blending_indices( self.dataset_index, self.dataset_sample_index, weights, num_datasets, self.size, torch.distributed.get_rank() == 0, ) print("> elapsed time for building blendable dataset indices: " "{:.2f} (sec)".format(time.time() - start_time)) def __len__(self): return self.size def __getitem__(self, idx): dataset_idx = self.dataset_index[idx] sample_idx = self.dataset_sample_index[idx] return self.datasets[dataset_idx][sample_idx]
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/__init__.py
examples/tutorial/sequence_parallel/data/datasets/__init__.py
from . import indexed_dataset
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/builder.py
examples/tutorial/sequence_parallel/data/datasets/builder.py
from colossalai.logging import get_dist_logger from .bert_dataset import BertDataset from .blendable_dataset import BlendableDataset from .dataset_utils import get_datasets_weights_and_num_samples, get_indexed_dataset_, get_train_valid_test_split_ DSET_TYPE_BERT = "standard_bert" DSET_TYPE_ICT = "ict" DSET_TYPE_T5 = "t5" DSET_TYPES = [DSET_TYPE_BERT, DSET_TYPE_ICT, DSET_TYPE_T5] def _build_train_valid_test_datasets( data_prefix, data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type="standard_bert", ): if dataset_type not in DSET_TYPES: raise ValueError("Invalid dataset_type: ", dataset_type) # Indexed dataset. indexed_dataset = get_indexed_dataset_(data_prefix, data_impl, skip_warmup) # Get start and end indices of train/valid/train into doc-idx # Note that doc-idx is designed to be num-docs + 1 so we can # easily iterate over it. total_num_of_documents = indexed_dataset.doc_idx.shape[0] - 1 splits = get_train_valid_test_split_(splits_string, total_num_of_documents) logger = get_dist_logger() # Print stats about the splits. logger.info("\n > dataset split:", ranks=[0]) def print_split_stats(name, index): start_index = indexed_dataset.doc_idx[splits[index]] end_index = indexed_dataset.doc_idx[splits[index + 1]] logger.info( "\n {}:".format(name) + "\n document indices in [{}, {}) total of {} documents".format( splits[index], splits[index + 1], splits[index + 1] - splits[index] ) + "\n sentence indices in [{}, {}) total of {} sentences".format( start_index, end_index, end_index - start_index ), ranks=[0], ) print_split_stats("train", 0) print_split_stats("validation", 1) print_split_stats("test", 2) def build_dataset(index, name): dataset = None if splits[index + 1] > splits[index]: # Get the pointer to the original doc-idx so we can set it later. doc_idx_ptr = indexed_dataset.get_doc_idx() # Slice the doc-idx start_index = splits[index] # Add +1 so we can index into the dataset to get the upper bound. end_index = splits[index + 1] + 1 # New doc_idx view. indexed_dataset.set_doc_idx(doc_idx_ptr[start_index:end_index]) # Build the dataset accordingly. kwargs = dict( name=name, data_prefix=data_prefix, num_epochs=None, max_num_samples=train_valid_test_num_samples[index], max_seq_length=max_seq_length, seed=seed, ) if dataset_type != DSET_TYPE_BERT: raise NotImplementedError("Only BERT dataset is supported") else: dataset = BertDataset( indexed_dataset=indexed_dataset, masked_lm_prob=masked_lm_prob, short_seq_prob=short_seq_prob, binary_head=binary_head, **kwargs, ) # Set the original pointer so dataset remains the main dataset. indexed_dataset.set_doc_idx(doc_idx_ptr) # Checks. assert indexed_dataset.doc_idx[0] == 0 assert indexed_dataset.doc_idx.shape[0] == (total_num_of_documents + 1) return dataset train_dataset = build_dataset(0, "train") valid_dataset = build_dataset(1, "valid") test_dataset = build_dataset(2, "test") return (train_dataset, valid_dataset, test_dataset) def build_train_valid_test_datasets( data_prefix, data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type="standard_bert", ): if len(data_prefix) == 1: return _build_train_valid_test_datasets( data_prefix[0], data_impl, splits_string, train_valid_test_num_samples, max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type=dataset_type, ) # Blending dataset. # Parse the values. output = get_datasets_weights_and_num_samples(data_prefix, train_valid_test_num_samples) prefixes, weights, datasets_train_valid_test_num_samples = output # Build individual datasets. train_datasets = [] valid_datasets = [] test_datasets = [] for i in range(len(prefixes)): train_ds, valid_ds, test_ds = _build_train_valid_test_datasets( prefixes[i], data_impl, splits_string, datasets_train_valid_test_num_samples[i], max_seq_length, masked_lm_prob, short_seq_prob, seed, skip_warmup, binary_head, dataset_type=dataset_type, ) if train_ds: train_datasets.append(train_ds) if valid_ds: valid_datasets.append(valid_ds) if test_ds: test_datasets.append(test_ds) # Blend. blending_train_dataset = None if train_datasets: blending_train_dataset = BlendableDataset(train_datasets, weights) blending_valid_dataset = None if valid_datasets: blending_valid_dataset = BlendableDataset(valid_datasets, weights) blending_test_dataset = None if test_datasets: blending_test_dataset = BlendableDataset(test_datasets, weights) return (blending_train_dataset, blending_valid_dataset, blending_test_dataset)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/indexed_dataset.py
examples/tutorial/sequence_parallel/data/datasets/indexed_dataset.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # copied from fairseq/fairseq/data/indexed_dataset.py # Removed IndexedRawTextDataset since it relied on Fairseq dictionary # other slight modifications to remove fairseq dependencies # Added document index to index file and made it accessible. # An empty sentence no longer separates documents. import os import shutil import struct from functools import lru_cache from itertools import accumulate import numpy as np import torch def __best_fitting_dtype(vocab_size=None): if vocab_size is not None and vocab_size < 65500: return np.uint16 else: return np.int32 def get_available_dataset_impl(): return ["lazy", "cached", "mmap"] def infer_dataset_impl(path): if IndexedDataset.exists(path): with open(index_file_path(path), "rb") as f: magic = f.read(8) if magic == IndexedDataset._HDR_MAGIC: return "cached" elif magic == MMapIndexedDataset.Index._HDR_MAGIC[:8]: return "mmap" else: return None else: print(f"Dataset does not exist: {path}") print("Path should be a basename that both .idx and .bin can be appended to get full filenames.") return None def make_builder(out_file, impl, vocab_size=None): if impl == "mmap": return MMapIndexedDatasetBuilder(out_file, dtype=__best_fitting_dtype(vocab_size)) else: return IndexedDatasetBuilder(out_file) def make_dataset(path, impl, skip_warmup=False): if not IndexedDataset.exists(path): print(f"Dataset does not exist: {path}") print("Path should be a basename that both .idx and .bin can be appended to get full filenames.") return None if impl == "infer": impl = infer_dataset_impl(path) if impl == "lazy" and IndexedDataset.exists(path): return IndexedDataset(path) elif impl == "cached" and IndexedDataset.exists(path): return IndexedCachedDataset(path) elif impl == "mmap" and MMapIndexedDataset.exists(path): return MMapIndexedDataset(path, skip_warmup) print(f"Unknown dataset implementation: {impl}") return None def dataset_exists(path, impl): if impl == "mmap": return MMapIndexedDataset.exists(path) else: return IndexedDataset.exists(path) def read_longs(f, n): a = np.empty(n, dtype=np.int64) f.readinto(a) return a def write_longs(f, a): f.write(np.array(a, dtype=np.int64)) dtypes = {1: np.uint8, 2: np.int8, 3: np.int16, 4: np.int32, 5: np.int64, 6: float, 7: np.double, 8: np.uint16} def code(dtype): for k in dtypes.keys(): if dtypes[k] == dtype: return k raise ValueError(dtype) def index_file_path(prefix_path): return prefix_path + ".idx" def data_file_path(prefix_path): return prefix_path + ".bin" def create_doc_idx(sizes): doc_idx = [0] for i, s in enumerate(sizes): if s == 0: doc_idx.append(i + 1) return doc_idx class IndexedDataset(torch.utils.data.Dataset): """Loader for IndexedDataset""" _HDR_MAGIC = b"TNTIDX\x00\x00" def __init__(self, path): super().__init__() self.path = path self.data_file = None self.read_index(path) def read_index(self, path): with open(index_file_path(path), "rb") as f: magic = f.read(8) assert magic == self._HDR_MAGIC, ( "Index file doesn't match expected format. " "Make sure that --dataset-impl is configured properly." ) version = f.read(8) assert struct.unpack("<Q", version) == (1,) code, self.element_size = struct.unpack("<QQ", f.read(16)) self.dtype = dtypes[code] self._len, self.s = struct.unpack("<QQ", f.read(16)) self.doc_count = struct.unpack("<Q", f.read(8)) self.dim_offsets = read_longs(f, self._len + 1) self.data_offsets = read_longs(f, self._len + 1) self.sizes = read_longs(f, self.s) self.doc_idx = read_longs(f, self.doc_count) def read_data(self, path): self.data_file = open(data_file_path(path), "rb", buffering=0) def check_index(self, i): if i < 0 or i >= self._len: raise IndexError("index out of range") def __del__(self): if self.data_file: self.data_file.close() # @lru_cache(maxsize=8) def __getitem__(self, idx): if not self.data_file: self.read_data(self.path) if isinstance(idx, int): i = idx self.check_index(i) tensor_size = self.sizes[self.dim_offsets[i] : self.dim_offsets[i + 1]] a = np.empty(tensor_size, dtype=self.dtype) self.data_file.seek(self.data_offsets[i] * self.element_size) self.data_file.readinto(a) return a elif isinstance(idx, slice): start, stop, step = idx.indices(len(self)) if step != 1: raise ValueError("Slices into indexed_dataset must be contiguous") sizes = self.sizes[self.dim_offsets[start] : self.dim_offsets[stop]] size = sum(sizes) a = np.empty(size, dtype=self.dtype) self.data_file.seek(self.data_offsets[start] * self.element_size) self.data_file.readinto(a) offsets = list(accumulate(sizes)) sents = np.split(a, offsets[:-1]) return sents def __len__(self): return self._len def num_tokens(self, index): return self.sizes[index] def size(self, index): return self.sizes[index] @staticmethod def exists(path): return os.path.exists(index_file_path(path)) and os.path.exists(data_file_path(path)) @property def supports_prefetch(self): return False # avoid prefetching to save memory class IndexedCachedDataset(IndexedDataset): def __init__(self, path): super().__init__(path) self.cache = None self.cache_index = {} @property def supports_prefetch(self): return True def prefetch(self, indices): if all(i in self.cache_index for i in indices): return if not self.data_file: self.read_data(self.path) indices = sorted(set(indices)) total_size = 0 for i in indices: total_size += self.data_offsets[i + 1] - self.data_offsets[i] self.cache = np.empty(total_size, dtype=self.dtype) ptx = 0 self.cache_index.clear() for i in indices: self.cache_index[i] = ptx size = self.data_offsets[i + 1] - self.data_offsets[i] a = self.cache[ptx : ptx + size] self.data_file.seek(self.data_offsets[i] * self.element_size) self.data_file.readinto(a) ptx += size if self.data_file: # close and delete data file after prefetch so we can pickle self.data_file.close() self.data_file = None # @lru_cache(maxsize=8) def __getitem__(self, idx): if isinstance(idx, int): i = idx self.check_index(i) tensor_size = self.sizes[self.dim_offsets[i] : self.dim_offsets[i + 1]] a = np.empty(tensor_size, dtype=self.dtype) ptx = self.cache_index[i] np.copyto(a, self.cache[ptx : ptx + a.size]) return a elif isinstance(idx, slice): # Hack just to make this work, can optimizer later if necessary sents = [] for i in range(*idx.indices(len(self))): sents.append(self[i]) return sents class IndexedDatasetBuilder(object): element_sizes = {np.uint8: 1, np.int8: 1, np.int16: 2, np.int32: 4, np.int64: 8, float: 4, np.double: 8} def __init__(self, out_file, dtype=np.int32): self.out_file = open(out_file, "wb") self.dtype = dtype self.data_offsets = [0] self.dim_offsets = [0] self.sizes = [] self.element_size = self.element_sizes[self.dtype] self.doc_idx = [0] def add_item(self, tensor): bytes = self.out_file.write(np.array(tensor.numpy(), dtype=self.dtype)) self.data_offsets.append(self.data_offsets[-1] + bytes / self.element_size) for s in tensor.size(): self.sizes.append(s) self.dim_offsets.append(self.dim_offsets[-1] + len(tensor.size())) def end_document(self): self.doc_idx.append(len(self.sizes)) def merge_file_(self, another_file): index = IndexedDataset(another_file) assert index.dtype == self.dtype begin = self.data_offsets[-1] for offset in index.data_offsets[1:]: self.data_offsets.append(begin + offset) self.sizes.extend(index.sizes) begin = self.dim_offsets[-1] for dim_offset in index.dim_offsets[1:]: self.dim_offsets.append(begin + dim_offset) with open(data_file_path(another_file), "rb") as f: while True: data = f.read(1024) if data: self.out_file.write(data) else: break def finalize(self, index_file): self.out_file.close() index = open(index_file, "wb") index.write(b"TNTIDX\x00\x00") index.write(struct.pack("<Q", 1)) index.write(struct.pack("<QQ", code(self.dtype), self.element_size)) index.write(struct.pack("<QQ", len(self.data_offsets) - 1, len(self.sizes))) index.write(struct.pack("<Q", len(self.doc_idx))) write_longs(index, self.dim_offsets) write_longs(index, self.data_offsets) write_longs(index, self.sizes) write_longs(index, self.doc_idx) index.close() def _warmup_mmap_file(path): with open(path, "rb") as stream: while stream.read(100 * 1024 * 1024): pass class MMapIndexedDataset(torch.utils.data.Dataset): class Index(object): _HDR_MAGIC = b"MMIDIDX\x00\x00" @classmethod def writer(cls, path, dtype): class _Writer(object): def __enter__(self): self._file = open(path, "wb") self._file.write(cls._HDR_MAGIC) self._file.write(struct.pack("<Q", 1)) self._file.write(struct.pack("<B", code(dtype))) return self @staticmethod def _get_pointers(sizes): dtype_size = dtype().itemsize address = 0 pointers = [] for size in sizes: pointers.append(address) address += size * dtype_size return pointers def write(self, sizes, doc_idx): pointers = self._get_pointers(sizes) self._file.write(struct.pack("<Q", len(sizes))) self._file.write(struct.pack("<Q", len(doc_idx))) sizes = np.array(sizes, dtype=np.int32) self._file.write(sizes.tobytes(order="C")) del sizes pointers = np.array(pointers, dtype=np.int64) self._file.write(pointers.tobytes(order="C")) del pointers doc_idx = np.array(doc_idx, dtype=np.int64) self._file.write(doc_idx.tobytes(order="C")) def __exit__(self, exc_type, exc_val, exc_tb): self._file.close() return _Writer() def __init__(self, path, skip_warmup=False): with open(path, "rb") as stream: magic_test = stream.read(9) assert self._HDR_MAGIC == magic_test, ( "Index file doesn't match expected format. " "Make sure that --dataset-impl is configured properly." ) version = struct.unpack("<Q", stream.read(8)) assert (1,) == version (dtype_code,) = struct.unpack("<B", stream.read(1)) self._dtype = dtypes[dtype_code] self._dtype_size = self._dtype().itemsize self._len = struct.unpack("<Q", stream.read(8))[0] self._doc_count = struct.unpack("<Q", stream.read(8))[0] offset = stream.tell() if not skip_warmup: print(" warming up index mmap file...") _warmup_mmap_file(path) self._bin_buffer_mmap = np.memmap(path, mode="r", order="C") self._bin_buffer = memoryview(self._bin_buffer_mmap) print(" reading sizes...") self._sizes = np.frombuffer(self._bin_buffer, dtype=np.int32, count=self._len, offset=offset) print(" reading pointers...") self._pointers = np.frombuffer( self._bin_buffer, dtype=np.int64, count=self._len, offset=offset + self._sizes.nbytes ) print(" reading document index...") self._doc_idx = np.frombuffer( self._bin_buffer, dtype=np.int64, count=self._doc_count, offset=offset + self._sizes.nbytes + self._pointers.nbytes, ) def __del__(self): self._bin_buffer_mmap._mmap.close() del self._bin_buffer_mmap @property def dtype(self): return self._dtype @property def sizes(self): return self._sizes @property def doc_idx(self): return self._doc_idx @lru_cache(maxsize=8) def __getitem__(self, i): return self._pointers[i], self._sizes[i] def __len__(self): return self._len def __init__(self, path, skip_warmup=False): super().__init__() self._path = None self._index = None self._bin_buffer = None self._do_init(path, skip_warmup) def __getstate__(self): return self._path def __setstate__(self, state): self._do_init(state) def _do_init(self, path, skip_warmup): self._path = path self._index = self.Index(index_file_path(self._path), skip_warmup) if not skip_warmup: print(" warming up data mmap file...") _warmup_mmap_file(data_file_path(self._path)) print(" creating numpy buffer of mmap...") self._bin_buffer_mmap = np.memmap(data_file_path(self._path), mode="r", order="C") print(" creating memory view of numpy buffer...") self._bin_buffer = memoryview(self._bin_buffer_mmap) def __del__(self): self._bin_buffer_mmap._mmap.close() del self._bin_buffer_mmap del self._index def __len__(self): return len(self._index) # @lru_cache(maxsize=8) def __getitem__(self, idx): if isinstance(idx, int): ptr, size = self._index[idx] np_array = np.frombuffer(self._bin_buffer, dtype=self._index.dtype, count=size, offset=ptr) return np_array elif isinstance(idx, slice): start, stop, step = idx.indices(len(self)) if step != 1: raise ValueError("Slices into indexed_dataset must be contiguous") ptr = self._index._pointers[start] sizes = self._index._sizes[idx] offsets = list(accumulate(sizes)) total_size = sum(sizes) np_array = np.frombuffer(self._bin_buffer, dtype=self._index.dtype, count=total_size, offset=ptr) sents = np.split(np_array, offsets[:-1]) return sents def get(self, idx, offset=0, length=None): """Retrieves a single item from the dataset with the option to only return a portion of the item. get(idx) is the same as [idx] but get() does not support slicing. """ ptr, size = self._index[idx] if length is None: length = size - offset ptr += offset * np.dtype(self._index.dtype).itemsize np_array = np.frombuffer(self._bin_buffer, dtype=self._index.dtype, count=length, offset=ptr) return np_array @property def sizes(self): return self._index.sizes @property def doc_idx(self): return self._index.doc_idx def get_doc_idx(self): return self._index._doc_idx def set_doc_idx(self, doc_idx_): self._index._doc_idx = doc_idx_ @property def supports_prefetch(self): return False @staticmethod def exists(path): return os.path.exists(index_file_path(path)) and os.path.exists(data_file_path(path)) class MMapIndexedDatasetBuilder(object): def __init__(self, out_file, dtype=np.int64): self._data_file = open(out_file, "wb") self._dtype = dtype self._sizes = [] self._doc_idx = [0] def add_item(self, tensor): np_array = np.array(tensor.numpy(), dtype=self._dtype) self._data_file.write(np_array.tobytes(order="C")) self._sizes.append(np_array.size) def end_document(self): self._doc_idx.append(len(self._sizes)) def merge_file_(self, another_file): # Concatenate index index = MMapIndexedDataset.Index(index_file_path(another_file)) assert index.dtype == self._dtype for size in index.sizes: self._sizes.append(size) # Concatenate data with open(data_file_path(another_file), "rb") as f: shutil.copyfileobj(f, self._data_file) def finalize(self, index_file): self._data_file.close() with MMapIndexedDataset.Index.writer(index_file, self._dtype) as index: index.write(self._sizes, self._doc_idx)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/sequence_parallel/data/datasets/test/test_indexed_dataset.py
examples/tutorial/sequence_parallel/data/datasets/test/test_indexed_dataset.py
# This file isn't really a formal automated test, it's just a place to # put some code used during development and manual testing of # indexed_dataset. import argparse import os import sys from megatron.data import indexed_dataset from megatron.tokenizer import build_tokenizer script_dir = os.path.dirname(os.path.realpath(__file__)) sys.path.append(os.path.join(script_dir, "../../../")) def test_indexed_dataset(args): ds = indexed_dataset.make_dataset(args.data, args.dataset_impl) tokenizer = build_tokenizer(args) print(len(ds.doc_idx)) print(len(ds)) print(ds.doc_idx[-1]) if ds.supports_prefetch: # just prefetch the whole thing in test (so assume it is small) ds.prefetch(range(len(ds))) if args.count > len(ds.doc_idx) - 1: args.count = len(ds.doc_idx) - 1 for i in range(args.count): start = ds.doc_idx[i] end = ds.doc_idx[i + 1] ids = ds[start:end] print(f"Document {i}:") print("--------------") for s in ids: assert len(s) > 0 l = s.data.tolist() text = tokenizer.detokenize(l) print(text) print("---") def test_indexed_dataset_get(args): ds = indexed_dataset.make_dataset(args.data, args.dataset_impl) build_tokenizer(args) size = ds.sizes[0] print(f"size: {size}") full = ds.get(0) print(full) # print(tokenizer.detokenize(full.data.tolist())) print("---") end = ds.get(0, offset=size - 10) print(end) # print(tokenizer.detokenize(end.data.tolist())) start = ds.get(0, length=10) print(start) # print(tokenizer.detokenize(start.data.tolist())) part = ds.get(0, offset=2, length=8) print(part) # print(tokenizer.detokenize(part.data.tolist())) # def test_albert_dataset(args): # # tokenizer = FullBertTokenizer(args.vocab, do_lower_case=True) # # idataset = indexed_dataset.make_dataset(args.data, args.dataset_impl) # # ds = AlbertDataset(idataset, tokenizer) # ds = AlbertDataset.from_paths(args.vocab, args.data, args.dataset_impl, # args.epochs, args.max_num_samples, # args.masked_lm_prob, args.seq_length, # args.short_seq_prob, args.seed) # truncated = 0 # total = 0 # for i, s in enumerate(ds): # ids = s['text'] # tokens = ds.tokenizer.convert_ids_to_tokens(ids) # print(tokens) # if i >= args.count-1: # exit() def main(): parser = argparse.ArgumentParser() parser.add_argument("--data", type=str, help="prefix to data files") parser.add_argument("--dataset-impl", type=str, default="infer", choices=["lazy", "cached", "mmap", "infer"]) parser.add_argument("--count", type=int, default=10, help="Number of samples/documents to print") group = parser.add_argument_group(title="tokenizer") group.add_argument( "--tokenizer-type", type=str, required=True, choices=["BertWordPieceLowerCase", "GPT2BPETokenizer"], help="What type of tokenizer to use.", ) group.add_argument("--vocab-file", type=str, default=None, help="Path to the vocab file") group.add_argument("--merge-file", type=str, default=None, help="Path to the BPE merge file (if necessary).") parser.add_argument("--epochs", type=int, default=5, help="Number of epochs to plan for") parser.add_argument("--max-num-samples", type=int, default=None, help="Maximum number of samples to plan for") parser.add_argument("--masked-lm-prob", type=float, default=0.15, help="probability of masking tokens") parser.add_argument("--seq-length", type=int, default=512, help="maximum sequence length") parser.add_argument("--short-seq-prob", type=float, default=0.1, help="probability of creating a short sequence") parser.add_argument("--seed", type=int, default=1234, help="random seed") args = parser.parse_args() args.rank = 0 args.make_vocab_size_divisible_by = 128 args.tensor_model_parallel_size = 1 if args.dataset_impl == "infer": args.dataset_impl = indexed_dataset.infer_dataset_impl(args.data) # test_albert_dataset(args) test_indexed_dataset_get(args) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_ckpt_batchsize_test.py
examples/tutorial/auto_parallel/auto_ckpt_batchsize_test.py
from copy import deepcopy from functools import partial import torch import torchvision.models as tm from bench_utils import bench, data_gen_resnet import colossalai from colossalai.auto_parallel.checkpoint import CheckpointSolverRotor from colossalai.fx import metainfo_trace, symbolic_trace from colossalai.testing import spawn def _benchmark(rank, world_size, port): """Auto activation checkpoint batchsize benchmark This benchmark test the through put of Resnet152 with our activation solver given the memory budget of 95% of maximum GPU memory, and with the batch size of [512, 1024, 2048], you could see that using auto activation checkpoint with optimality guarantee, we might be able to find better batch size for the model, as larger batch size means that we are able to use larger portion of GPU FLOPS, while recomputation scheduling with our solver only result in minor performance drop. So at last we might be able to find better training batch size for our model (combine with large batch training optimizer such as LAMB). """ colossalai.launch(rank=rank, world_size=world_size, host="localhost", port=port, backend="nccl") model = tm.resnet152() gm = symbolic_trace(model) raw_graph = deepcopy(gm.graph) peak_mems, through_puts, batch_sizes = [], [], [512, 1024, 2048] for batch_size in batch_sizes: batch_size = int(batch_size) gm = metainfo_trace(gm, torch.empty(batch_size, 3, 224, 224, device="meta")) solver = CheckpointSolverRotor(gm.graph, free_memory=torch.cuda.mem_get_info()[0] * 0.95) gm.graph = solver.solve() peak_mem, step_time = bench( gm, torch.nn.CrossEntropyLoss(), partial(data_gen_resnet, batch_size=batch_size, shape=(3, 224, 224)), num_steps=5, ) peak_mems.append(peak_mem) through_puts.append(batch_size / step_time * 1.0e3) gm.graph = deepcopy(raw_graph) # print results print("===============benchmark summary================") for batch_size, peak_mem, through_put in zip(batch_sizes, peak_mems, through_puts): print(f"batch_size: {int(batch_size)}, peak memory: {peak_mem:.3f} MB, through put: {through_put:.3f} images/s") def auto_activation_checkpoint_batchsize_benchmark(): spawn(_benchmark, 1) if __name__ == "__main__": auto_activation_checkpoint_batchsize_benchmark()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/setup.py
examples/tutorial/auto_parallel/setup.py
from setuptools import find_packages, setup setup( name="auto_parallel", version="0.0.1", description="", packages=find_packages(), install_requires=[ "torch", "numpy", "tqdm", ], )
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/bench_utils.py
examples/tutorial/auto_parallel/bench_utils.py
import time from copy import deepcopy from typing import Callable, Tuple import numpy as np import torch import torch.nn as nn from transformers import GPT2Config, GPT2LMHeadModel from colossalai.auto_parallel.checkpoint import CheckpointSolverRotor from colossalai.fx import metainfo_trace def bench( gm: torch.fx.GraphModule, criterion: torch.nn.Module, data_gen: Callable, num_steps: int = 5 ) -> Tuple[int, int]: """Benchmarking a given graph module Args: gm (torch.fx.GraphModule): The graph module to benchmark. criterion (torch.nn.Module): Loss function. data_gen (Callable): Data generator. num_steps (int, optional): Number of test steps. Defaults to 5. Returns: Tuple[int, int]: peak memory in MB and step time in MS. """ gm.train() gm.cuda() step_time = float("inf") torch.cuda.synchronize() torch.cuda.empty_cache() torch.cuda.reset_peak_memory_stats() cached = torch.cuda.max_memory_allocated(device="cuda") try: for _ in range(num_steps): args, label = data_gen() output, loss = None, None torch.cuda.synchronize(device="cuda") start = time.time() output = gm(*args) loss = criterion(output, label) loss.backward() torch.cuda.synchronize(device="cuda") step_time = min(step_time, time.time() - start) for child in gm.children(): for param in child.parameters(): param.grad = None del args, label, output, loss except: del args, label, output, loss gm.to("cpu") torch.cuda.empty_cache() peak_mem = (torch.cuda.max_memory_allocated(device="cuda") - cached) / 1024**2 return peak_mem, step_time * 1.0e3 def bench_rotor( gm: torch.fx.GraphModule, criterion: torch.nn.Module, data_gen: Callable, num_steps: int = 5, sample_points: int = 20, free_memory: int = torch.cuda.mem_get_info()[0], start_factor: int = 4, ) -> Tuple[np.array, list, list]: """Auto Checkpoint Rotor Algorithm benchmarking Benchmarks the Auto Checkpoint Rotor Algorithm for a given graph module and data. Args: gm (torch.fx.GraphModule): The graph module to benchmark. criterion (torch.nn.Module): Loss function. data_gen (Callable): Data generator. num_steps (int, optional): Number of test steps. Defaults to 5. sample_points (int, optional): Number of sample points. Defaults to 20. free_memory (int, optional): Max memory budget in Byte. Defaults to torch.cuda.mem_get_info()[0]. start_factor (int, optional): Start memory budget factor for benchmark, the start memory budget will be free_memory / start_factor. Defaults to 4. Returns: Tuple[np.array, list, list]: return budgets vector (MB), peak memory vector (MB), step time vector (MS). """ peak_hist, step_hist = [], [] raw_graph = deepcopy(gm.graph) for budget in np.linspace(free_memory // start_factor, free_memory, sample_points): gm = metainfo_trace(gm, *data_gen()[0]) solver = CheckpointSolverRotor(gm.graph, free_memory=budget) try: gm.graph = solver.solve(verbose=False) peak_memory, step_time = bench(gm, criterion, data_gen, num_steps=num_steps) except: peak_memory, step_time = budget / 1024**2, float("inf") peak_hist.append(peak_memory) step_hist.append(step_time) gm.graph = deepcopy(raw_graph) return np.linspace(free_memory // start_factor, free_memory, sample_points) / 1024**2, peak_hist, step_hist class GPTLMModel(nn.Module): """ GPT Model """ def __init__( self, hidden_size=768, num_layers=12, num_attention_heads=12, max_seq_len=1024, vocab_size=50257, checkpoint=False, ): super().__init__() self.checkpoint = checkpoint self.model = GPT2LMHeadModel( GPT2Config( n_embd=hidden_size, n_layer=num_layers, n_head=num_attention_heads, n_positions=max_seq_len, n_ctx=max_seq_len, vocab_size=vocab_size, ) ) if checkpoint: self.model.gradient_checkpointing_enable() def forward(self, input_ids, attention_mask): # Only return lm_logits return self.model(input_ids=input_ids, attention_mask=attention_mask, use_cache=not self.checkpoint)[0] class GPTLMLoss(nn.Module): """ GPT Loss """ def __init__(self): super().__init__() self.loss_fn = nn.CrossEntropyLoss() def forward(self, logits, labels): shift_logits = logits[..., :-1, :].contiguous() shift_labels = labels[..., 1:].contiguous() # Flatten the tokens return self.loss_fn(shift_logits.view(-1, shift_logits.size(-1)), shift_labels.view(-1)) def gpt2_medium(checkpoint=False): return GPTLMModel(hidden_size=1024, num_layers=24, num_attention_heads=16, checkpoint=checkpoint) def gpt2_xl(checkpoint=False): return GPTLMModel(hidden_size=1600, num_layers=48, num_attention_heads=32, checkpoint=checkpoint) def gpt2_6b(checkpoint=False): return GPTLMModel(hidden_size=4096, num_layers=30, num_attention_heads=16, checkpoint=checkpoint) def data_gen_gpt2(batch_size, seq_len, vocab_size, device="cuda:0"): """ Generate random data for gpt2 benchmarking """ input_ids = torch.randint(0, vocab_size, (batch_size, seq_len), device=device) attention_mask = torch.ones_like(input_ids, device=device) return (input_ids, attention_mask), attention_mask def data_gen_resnet(batch_size, shape, device="cuda:0"): """ Generate random data for resnet benchmarking """ data = torch.empty(batch_size, *shape, device=device) label = torch.empty(batch_size, dtype=torch.long, device=device).random_(1000) return (data,), label
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_parallel_with_resnet.py
examples/tutorial/auto_parallel/auto_parallel_with_resnet.py
import torch from torchvision.models import resnet50 from tqdm import tqdm import colossalai from colossalai.auto_parallel.tensor_shard.initialize import initialize_model from colossalai.device.device_mesh import DeviceMesh from colossalai.legacy.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn.lr_scheduler import CosineAnnealingLR def synthesize_data(): img = torch.rand(gpc.config.BATCH_SIZE, 3, 32, 32) label = torch.randint(low=0, high=10, size=(gpc.config.BATCH_SIZE,)) return img, label def main(): colossalai.legacy.launch_from_torch(config="./config.py") logger = get_dist_logger() # trace the model with meta data model = resnet50(num_classes=10).cuda() input_sample = {"x": torch.rand([gpc.config.BATCH_SIZE * torch.distributed.get_world_size(), 3, 32, 32]).to("meta")} device_mesh = DeviceMesh(physical_mesh_id=torch.tensor([0, 1, 2, 3]), mesh_shape=[2, 2], init_process_group=True) model, solution = initialize_model(model, input_sample, device_mesh=device_mesh, return_solution=True) if gpc.get_global_rank() == 0: for node_strategy in solution: print(node_strategy) # build criterion criterion = torch.nn.CrossEntropyLoss() # optimizer optimizer = torch.optim.SGD(model.parameters(), lr=0.1, momentum=0.9, weight_decay=5e-4) # lr_scheduler lr_scheduler = CosineAnnealingLR(optimizer, total_steps=gpc.config.NUM_EPOCHS) for epoch in range(gpc.config.NUM_EPOCHS): model.train() # if we use synthetic data # we assume it only has 10 steps per epoch num_steps = range(10) progress = tqdm(num_steps) for _ in progress: # generate fake data img, label = synthesize_data() img = img.cuda() label = label.cuda() optimizer.zero_grad() output = model(img) train_loss = criterion(output, label) train_loss.backward(train_loss) torch.cuda.synchronize() optimizer.step() lr_scheduler.step() # run evaluation model.eval() correct = 0 total = 0 # if we use synthetic data # we assume it only has 10 steps for evaluation num_steps = range(10) progress = tqdm(num_steps) for _ in progress: # generate fake data img, label = synthesize_data() img = img.cuda() label = label.cuda() with torch.no_grad(): output = model(img) test_loss = criterion(output, label) pred = torch.argmax(output, dim=-1) correct += torch.sum(pred == label) total += img.size(0) logger.info( f"Epoch {epoch} - train loss: {train_loss:.5}, test loss: {test_loss:.5}, acc: {correct / total:.5}, lr: {lr_scheduler.get_last_lr()[0]:.5g}", ranks=[0], ) if __name__ == "__main__": main()
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/auto_ckpt_solver_test.py
examples/tutorial/auto_parallel/auto_ckpt_solver_test.py
from argparse import ArgumentParser from functools import partial import matplotlib.pyplot as plt import torch import torchvision.models as tm from bench_utils import GPTLMLoss, bench_rotor, data_gen_gpt2, data_gen_resnet, gpt2_medium import colossalai from colossalai.fx import metainfo_trace, symbolic_trace from colossalai.testing import spawn def _benchmark(rank, world_size, port, args): """ Auto activation checkpoint solver benchmark, we provide benchmark on two models: gpt2_medium and resnet50. The benchmark will sample in a range of memory budget for each model and output the benchmark summary and data visualization of peak memory vs. budget memory and relative step time vs. peak memory. """ colossalai.launch(rank=rank, world_size=world_size, host="localhost", port=port, backend="nccl") if args.model == "resnet50": model = tm.resnet50() data_gen = partial(data_gen_resnet, batch_size=128, shape=(3, 224, 224)) gm = symbolic_trace(model) gm = metainfo_trace(gm, torch.empty(128, 3, 224, 224, device="meta")) loss = torch.nn.CrossEntropyLoss() else: model = gpt2_medium() data_gen = partial(data_gen_gpt2, batch_size=8, seq_len=1024, vocab_size=50257) data, mask = data_gen(device="meta")[0] gm = symbolic_trace(model, meta_args={"input_ids": data, "attention_mask": mask}) gm = metainfo_trace(gm, data, mask) loss = GPTLMLoss() free_memory = 11000 * 1024**2 if args.model == "resnet50" else 56000 * 1024**2 start_factor = 4 if args.model == "resnet50" else 10 # trace and benchmark budgets, peak_hist, step_hist = bench_rotor( gm, loss, data_gen, num_steps=5, sample_points=15, free_memory=free_memory, start_factor=start_factor ) # print summary print("==============benchmark summary==============") for budget, peak, step in zip(budgets, peak_hist, step_hist): print(f"memory budget: {budget:.3f} MB, peak memory: {peak:.3f} MB, step time: {step:.3f} MS") # plot valid results fig, axs = plt.subplots(1, 2, figsize=(16, 8)) valid_idx = step_hist.index(next(step for step in step_hist if step != float("inf"))) # plot peak memory vs. budget memory axs[0].plot(budgets[valid_idx:], peak_hist[valid_idx:]) axs[0].plot([budgets[valid_idx], budgets[-1]], [budgets[valid_idx], budgets[-1]], linestyle="--") axs[0].set_xlabel("Budget Memory (MB)") axs[0].set_ylabel("Peak Memory (MB)") axs[0].set_title("Peak Memory vs. Budget Memory") # plot relative step time vs. budget memory axs[1].plot(peak_hist[valid_idx:], [step_time / step_hist[-1] for step_time in step_hist[valid_idx:]]) axs[1].plot([peak_hist[valid_idx], peak_hist[-1]], [1.0, 1.0], linestyle="--") axs[1].set_xlabel("Peak Memory (MB)") axs[1].set_ylabel("Relative Step Time") axs[1].set_title("Step Time vs. Peak Memory") axs[1].set_ylim(0.8, 1.5) # save plot fig.savefig(f"{args.model}_benchmark.png") def auto_activation_checkpoint_benchmark(args): world_size = 1 spawn(_benchmark, world_size, args=args) if __name__ == "__main__": parser = ArgumentParser("Auto Activation Checkpoint Solver Benchmark") parser.add_argument("--model", type=str, default="gpt2", choices=["gpt2", "resnet50"]) args = parser.parse_args() auto_activation_checkpoint_benchmark(args)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/tutorial/auto_parallel/config.py
examples/tutorial/auto_parallel/config.py
BATCH_SIZE = 32 NUM_EPOCHS = 2
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_context_attn_unpad.py
examples/inference/benchmark_ops/benchmark_context_attn_unpad.py
import torch from transformers.modeling_attn_mask_utils import AttentionMaskConverter from colossalai.inference.modeling.layers.attention import PagedAttention from colossalai.kernel.triton import context_attention_unpadded from colossalai.utils import get_current_device from tests.test_infer.test_kernels.triton.kernel_utils import generate_caches_and_block_tables_v2, torch_attn_ref try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") HEAD_DIM = 32 BATCH = 16 BLOCK_SIZE = 32 SAME_LEN = True WARM_UPS = 10 REPS = 100 configs = [ triton.testing.Benchmark( x_names=["KV_LEN"], x_vals=[2**i for i in range(8, 13)], # x_vals=[x for x in range(256, 8192, 256)], line_arg="provider", line_vals=["torch", "triton", "triton_new_klayout"], line_names=["Torch", "Triton", "Triton_new_klayout"], styles=[("red", "-"), ("blue", "-"), ("green", "-")], ylabel="ms", plot_name=f"context_attn-block_size-{BLOCK_SIZE}-batch{BATCH}", args={"bsz": BATCH, "block_size": BLOCK_SIZE, "same_context_len": SAME_LEN, "kv_group_num": 1}, ) ] @triton.testing.perf_report(configs) def bench_kernel( bsz, KV_LEN, provider, block_size: int, kv_group_num: int, same_context_len: bool, ): num_attn_heads = 16 max_num_blocks_per_seq = triton.cdiv(KV_LEN, block_size) max_seq_len = block_size * max_num_blocks_per_seq num_kv_heads = num_attn_heads // kv_group_num assert isinstance(num_kv_heads, int) and num_kv_heads > 0, "Invalid number of kv heads." dtype = torch.float16 device = get_current_device() if same_context_len: context_lengths = torch.tensor([max_seq_len for _ in range(bsz)], dtype=torch.int32, device=device) else: context_lengths = torch.randint(low=1, high=max_seq_len, size=(bsz,), dtype=torch.int32, device=device) num_tokens = torch.sum(context_lengths).item() qkv_size = (num_tokens, num_attn_heads + 2 * num_kv_heads, HEAD_DIM) qkv_unpad = torch.empty(size=qkv_size, dtype=dtype, device=device).normal_(mean=0.0, std=0.5) q_unpad, k_unpad, v_unpad = torch.split(qkv_unpad, [num_attn_heads, num_kv_heads, num_kv_heads], dim=-2) q_unpad = q_unpad.contiguous() k_cache_ref, v_cache_ref, block_tables = generate_caches_and_block_tables_v2( k_unpad, v_unpad, context_lengths, bsz, max_num_blocks_per_seq, block_size, dtype, device ) block_tables = block_tables.to(device=device) quantiles = [0.5, 0.2, 0.8] if provider == "torch": q_padded = PagedAttention.pad_and_reshape(q_unpad, context_lengths, max_seq_len, num_attn_heads, HEAD_DIM) k_padded = PagedAttention.pad_and_reshape(k_unpad, context_lengths, max_seq_len, num_kv_heads, HEAD_DIM) v_padded = PagedAttention.pad_and_reshape(v_unpad, context_lengths, max_seq_len, num_kv_heads, HEAD_DIM) q_padded, k_padded, v_padded = ( q_padded.to(device=device), k_padded.to(device=device), v_padded.to(device=device), ) q_padded = q_padded.transpose(1, 2) k_padded = PagedAttention.repeat_kv(k_padded.transpose(1, 2), kv_group_num) v_padded = PagedAttention.repeat_kv(v_padded.transpose(1, 2), kv_group_num) # This benchmark ignores the padding mask. *Only* use the-same-length inputs for benchmarkings attn_mask = AttentionMaskConverter._make_causal_mask( (bsz, max_seq_len), q_padded.dtype, q_padded.device, past_key_values_length=0 ) attn_mask = attn_mask.to(device=q_padded.device) fn = lambda: torch_attn_ref( q_padded, k_padded, v_padded, attn_mask, bsz, max_seq_len, max_seq_len, num_attn_heads, num_kv_heads, HEAD_DIM, ) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) elif provider == "triton": k_cache_triton = torch.zeros_like(k_cache_ref) v_cache_triton = torch.zeros_like(v_cache_ref) fn = lambda: context_attention_unpadded( q_unpad, k_unpad, v_unpad, k_cache_triton, v_cache_triton, context_lengths, block_tables, block_size ) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) elif provider == "triton_new_klayout": # NOTE New kcache layout (num_blocks, num_kv_heads, head_dim // x, block_size, x) # to be applied around the cuda and triton kernels. # Here we want to make sure it does not cause downgrade in performance. x = 16 // torch.tensor([], dtype=dtype).element_size() k_cache_shape = (bsz * max_num_blocks_per_seq, num_kv_heads, HEAD_DIM // x, block_size, x) k_cache_triton = torch.zeros(size=k_cache_shape, dtype=dtype, device=device) v_cache_triton = torch.zeros_like(v_cache_ref) fn = lambda: context_attention_unpadded( q_unpad, k_unpad, v_unpad, k_cache_triton, v_cache_triton, context_lengths, block_tables, block_size, use_new_kcache_layout=True, ) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) return ms, min_ms, max_ms if __name__ == "__main__": bench_kernel.run(save_path=".", print_data=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_fused_rotary_embdding_unpad.py
examples/inference/benchmark_ops/benchmark_fused_rotary_embdding_unpad.py
import torch from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import copy_kv_to_blocked_cache, decoding_fused_rotary_embedding, rotary_embedding from tests.test_infer.test_kernels.triton.kernel_utils import ( mock_alloc_block_table_and_kvcache_v2, mock_alloc_block_table_and_kvcache_v3, mock_alloc_single_token, ) inference_ops = InferenceOpsLoader().load() try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") BATCH = 16 configs = [ triton.testing.Benchmark( x_names=["num_tokens"], x_vals=[2**i for i in range(4, 11)], line_arg="provider", line_vals=[ "triton_rotary_emb_func", "triton_fused_rotary_emb_func", "triton_fused_rotary_emb_func_new_kcache_layout", "cuda_rotary_emb_func", "cuda_fused_rotary_emb_func", ], line_names=[ "triton_rotary_emb_func", "triton_fused_rotary_emb_func", "triton_fused_rotary_emb_func(new layout)", "cuda_rotary_emb_func", "cuda_fused_rotary_emb_func", ], styles=[("red", "-"), ("blue", "-"), ("purple", "-"), ("green", "-"), ("yellow", "-")], ylabel="ms", plot_name=f"rotary_emb-batch-{BATCH}", args={"num_kv_heads": 16}, ) ] @triton.testing.perf_report(configs) def benchmark_rotary_emb( provider: str, num_tokens: int, num_kv_heads: int, ): BATCH_SIZE = 16 SEQ_LEN = num_tokens // BATCH_SIZE max_num_blocks_per_seq = 8 block_size = 64 warmup = 10 rep = 100 head_dim = 4096 dtype = torch.float16 q_shape = (num_tokens, num_kv_heads, head_dim) q = -2.3 + 0.5 * torch.randn(q_shape, dtype=dtype, device="cuda") k_shape = (num_tokens, num_kv_heads, head_dim) k = -2.3 + 0.5 * torch.randn(k_shape, dtype=dtype, device="cuda") v = -2.3 + 0.5 * torch.randn(k_shape, dtype=dtype, device="cuda") cos_shape = (num_tokens, head_dim // 2) cos = -1.2 + 0.5 * torch.randn(cos_shape, dtype=dtype, device="cuda") sin = -2.0 + 0.5 * torch.randn(cos_shape, dtype=dtype, device="cuda") cache_shape = (BATCH_SIZE * max_num_blocks_per_seq, num_kv_heads, block_size, head_dim) k_cache = torch.zeros(size=cache_shape, dtype=dtype, device="cuda") v_cache = torch.zeros(size=cache_shape, dtype=dtype, device="cuda") x = 16 // torch.tensor([], dtype=dtype).element_size() new_cache_shape = (BATCH_SIZE * max_num_blocks_per_seq, num_kv_heads, head_dim // x, block_size, x) new_k_cache = torch.zeros(size=new_cache_shape, dtype=dtype, device="cuda") past_kv_seq_lengths = torch.tensor([SEQ_LEN - 1 for _ in range(BATCH_SIZE)], dtype=torch.int32, device="cuda") block_tables = mock_alloc_block_table_and_kvcache_v2( k, v, k_cache, v_cache, past_kv_seq_lengths, BATCH_SIZE, max_num_blocks_per_seq, block_size ) _ = mock_alloc_block_table_and_kvcache_v3( k, v, new_k_cache, v_cache, past_kv_seq_lengths, BATCH_SIZE, max_num_blocks_per_seq, block_size ) new_k = torch.randn((BATCH_SIZE, num_kv_heads, head_dim), dtype=dtype, device="cuda") new_q = torch.randn_like(new_k) new_v = torch.randn_like(new_k) mock_alloc_single_token(block_tables, past_kv_seq_lengths, block_size) kv_seq_lengths = past_kv_seq_lengths + 1 block_tables = block_tables.to(device="cuda") quantiles = [0.5, 0.2, 0.8] if provider == "triton_rotary_emb_func": fn = lambda: [ rotary_embedding(new_q, new_k, cos, sin), copy_kv_to_blocked_cache( new_k, new_v, k_cache, v_cache, kv_lengths=kv_seq_lengths, block_tables=block_tables ), ] elif provider == "triton_fused_rotary_emb_func": fn = lambda: decoding_fused_rotary_embedding( new_q, new_k, new_v, cos, sin, k_cache, v_cache, block_tables, kv_seq_lengths ) elif provider == "triton_fused_rotary_emb_func_new_kcache_layout": x = 16 // torch.tensor([], dtype=dtype).element_size() kcache_shape = (BATCH_SIZE * max_num_blocks_per_seq, num_kv_heads, head_dim // x, block_size, x) k_cache = torch.zeros(size=kcache_shape, dtype=dtype, device="cuda") block_tables = mock_alloc_block_table_and_kvcache_v3( k, v, k_cache, v_cache, past_kv_seq_lengths, BATCH_SIZE, max_num_blocks_per_seq, block_size ) mock_alloc_single_token(block_tables, past_kv_seq_lengths, block_size) block_tables = block_tables.to(device="cuda") fn = lambda: decoding_fused_rotary_embedding( new_q, new_k, new_v, cos, sin, k_cache, v_cache, block_tables, kv_seq_lengths, use_new_kcache_layout=True ) elif provider == "cuda_rotary_emb_func": fn = lambda: [ inference_ops.rotary_embedding(new_q, new_k, cos, sin, True), inference_ops.decode_kv_cache_memcpy(new_k, new_v, new_k_cache, v_cache, kv_seq_lengths, block_tables), ] elif provider == "cuda_fused_rotary_emb_func": fn = lambda: inference_ops.rotary_embedding_and_cache_copy( new_q, new_k, new_v, cos, sin, new_k_cache, v_cache, kv_seq_lengths, block_tables, True ) else: raise ValueError("Undefined provider") ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=warmup, rep=rep, quantiles=quantiles) return ms, min_ms, max_ms if __name__ == "__main__": benchmark_rotary_emb.run(save_path=".", print_data=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_decoding_attn.py
examples/inference/benchmark_ops/benchmark_decoding_attn.py
import torch from colossalai.kernel.triton import flash_decoding_attention from colossalai.utils import get_current_device from tests.test_infer.test_kernels.triton.kernel_utils import ( convert_kv_unpad_to_padded, create_attention_mask, generate_caches_and_block_tables_v2, generate_caches_and_block_tables_v3, torch_attn_ref, ) from tests.test_infer.test_kernels.triton.test_decoding_attn import prepare_data try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") Q_LEN = 1 HEAD_DIM = 128 BATCH = 16 BLOCK_SIZE = 32 SAME_LEN = True WARM_UPS = 10 REPS = 100 configs = [ triton.testing.Benchmark( x_names=["KV_LEN"], x_vals=[2**i for i in range(8, 14)], # x_vals=[x for x in range(256, 8192, 256)], line_arg="provider", line_vals=["torch", "triton", "triton_new_kcache_layout"], line_names=["Torch", "Triton", "Triton New KCache Layout"], styles=[("red", "-"), ("blue", "-"), ("yellow", "-")], ylabel="ms", plot_name=f"decoding-block_size-{BLOCK_SIZE}-batch{BATCH}", args={"bsz": BATCH, "block_size": BLOCK_SIZE, "same_context_len": SAME_LEN, "kv_group_num": 1}, ) ] @triton.testing.perf_report(configs) def bench_kernel( bsz, KV_LEN, provider, block_size: int, kv_group_num: int, same_context_len: bool, ): num_attn_heads = 16 max_num_blocks_per_seq = triton.cdiv(KV_LEN, block_size) max_seq_len = block_size * max_num_blocks_per_seq num_kv_heads = num_attn_heads // kv_group_num assert isinstance(num_kv_heads, int) and num_kv_heads > 0, "Invalid number of kv heads." block_size * max_num_blocks_per_seq dtype = torch.float16 device = get_current_device() q, k_unpad, v_unpad, kv_lengths = prepare_data( bsz, num_attn_heads, num_kv_heads, HEAD_DIM, same_context_len, Q_LEN, max_seq_len, dtype, device ) max_seq_len_in_b = kv_lengths.max().item() # for random lengths # the maximum block length splitted on kv should be the kv cache block size kv_max_split_num = (max_seq_len_in_b + block_size - 1) // block_size sm_scale = 1.0 / (HEAD_DIM**0.5) output = torch.empty((bsz, num_attn_heads, HEAD_DIM), dtype=dtype, device=device) mid_output = torch.empty( size=(bsz, num_attn_heads, kv_max_split_num, HEAD_DIM), dtype=torch.float32, device=q.device ) mid_output_lse = torch.empty(size=(bsz, num_attn_heads, kv_max_split_num), dtype=torch.float32, device=q.device) quantiles = [0.5, 0.2, 0.8] if provider == "torch": k_torch = convert_kv_unpad_to_padded(k_unpad, kv_lengths, bsz, max_seq_len_in_b) v_torch = convert_kv_unpad_to_padded(v_unpad, kv_lengths, bsz, max_seq_len_in_b) torch_padding_mask = create_attention_mask(kv_lengths, bsz, Q_LEN, max_seq_len_in_b, q.device) fn = lambda: torch_attn_ref( q, k_torch, v_torch, torch_padding_mask, bsz, Q_LEN, max_seq_len_in_b, num_attn_heads, num_kv_heads, HEAD_DIM, ) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) elif provider == "triton": k_cache, v_cache, block_tables = generate_caches_and_block_tables_v2( k_unpad, v_unpad, kv_lengths, bsz, max_num_blocks_per_seq, block_size, dtype, device ) block_tables = block_tables.to(device=device) fn = lambda: flash_decoding_attention( # Here we use q.squeeze(2) because we hide the q_len dimension (which is equivalent to 1), # refer to attention forward in modeling. q.squeeze(2), k_cache, v_cache, kv_lengths, block_tables, block_size, max_seq_len_in_b, output, mid_output, mid_output_lse, sm_scale=sm_scale, kv_group_num=kv_group_num, ) # [bsz, 1, num_heads, head_dim] ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) elif provider == "triton_new_kcache_layout": k_cache, v_cache, block_tables = generate_caches_and_block_tables_v3( k_unpad, v_unpad, kv_lengths, bsz, max_num_blocks_per_seq, block_size, dtype, device ) block_tables = block_tables.to(device=device) fn = lambda: flash_decoding_attention( # Here we use q.squeeze(2) because we hide the q_len dimension (which is equivalent to 1), # refer to attention forward in modeling. q.squeeze(2), k_cache, v_cache, kv_lengths, block_tables, block_size, max_seq_len_in_b, output, mid_output, mid_output_lse, sm_scale=sm_scale, kv_group_num=kv_group_num, use_new_kcache_layout=True, ) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) return ms, min_ms, max_ms if __name__ == "__main__": bench_kernel.run(save_path=".", print_data=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_kv_cache_memcopy.py
examples/inference/benchmark_ops/benchmark_kv_cache_memcopy.py
import torch from colossalai.inference.modeling.layers.attention import copy_to_cache from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import copy_kv_to_blocked_cache from colossalai.utils import get_current_device from tests.test_infer.test_kernels.cuda.test_kv_cache_memcpy import prepare_data as prepare_data_new_kcache_layout from tests.test_infer.test_kernels.triton.test_kvcache_copy import prepare_data try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") inference_ops = InferenceOpsLoader().load() HEAD_DIM = 128 BATCH = 16 BLOCK_SIZE = 32 SAME_LEN = True WARM_UPS = 10 REPS = 100 configs = [ triton.testing.Benchmark( x_names=["KV_SEQ_LEN"], x_vals=[2**i for i in range(8, 13)], line_arg="provider", line_vals=["torch_copy_func", "triton_copy_func", "triton_new_kcache_layout", "cuda_copy_func"], line_names=["torch_copy_func", "triton_copy_func", "triton_new_kcache_layout", "cuda_copy_func"], styles=[("red", "-"), ("blue", "-"), ("yellow", "-"), ("green", "-")], ylabel="ms", plot_name=f"kvcache_copy_decoding_stage-batch-{BATCH}", args={"bsz": BATCH, "block_size": 16, "max_seq_len": 8192, "num_kv_heads": 16, "same_context_len": True}, ) ] @triton.testing.perf_report(configs) def benchmark_kvcache_copy( provider: str, bsz: int, block_size: int, max_seq_len: int, KV_SEQ_LEN: int, # maximum past kv length (unequal context lens in batch) or past kv len (equal context lens) num_kv_heads: int, same_context_len: bool, ): dtype = torch.float16 device = get_current_device() assert KV_SEQ_LEN <= max_seq_len, "Assigned maximum kv length must be smaller or equal to maximum seq len" new_k, new_v, k_cache, v_cache, context_lengths, block_tables = prepare_data( bsz, num_kv_heads, HEAD_DIM, block_size, max_seq_len // block_size, same_context_len, KV_SEQ_LEN, device=device, dtype=dtype, ) quantiles = [0.5, 0.2, 0.8] if provider == "torch_copy_func": fn = lambda: copy_to_cache(new_k, k_cache, lengths=context_lengths, block_tables=block_tables, type="decoding") elif provider == "triton_copy_func": fn = lambda: copy_kv_to_blocked_cache(new_k, new_v, k_cache, v_cache, context_lengths, block_tables) elif provider == "triton_new_kcache_layout": # NOTE New kcache layout (num_blocks, num_kv_heads, head_dim // x, block_size, x) to be applied x = 16 // torch.tensor([], dtype=dtype).element_size() k_cache_shape = (bsz * max_seq_len // block_size, num_kv_heads, HEAD_DIM // x, block_size, x) k_cache = torch.zeros(size=k_cache_shape, dtype=dtype, device=device) # update k_cache layout fn = lambda: copy_kv_to_blocked_cache( new_k, new_v, k_cache, v_cache, context_lengths, block_tables, use_new_kcache_layout=True ) elif provider == "cuda_copy_func": _, _, k_cache, _, _, _, _, _, _ = prepare_data_new_kcache_layout( bsz, num_kv_heads, block_size, max_seq_len // block_size, context_lengths - 1, device, dtype ) new_k = new_k.squeeze(1) if new_k.dim() == 4 else new_k new_v = new_v.squeeze(1) if new_v.dim() == 4 else new_v fn = lambda: inference_ops.decode_kv_cache_memcpy(new_k, new_v, k_cache, v_cache, context_lengths, block_tables) ms, min_ms, max_ms = triton.testing.do_bench(fn, warmup=WARM_UPS, rep=REPS, quantiles=quantiles) return ms, min_ms, max_ms if __name__ == "__main__": benchmark_kvcache_copy.run(save_path=".", print_data=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false
hpcaitech/ColossalAI
https://github.com/hpcaitech/ColossalAI/blob/b1915d2889543949eb5b610241f1515c73df5059/examples/inference/benchmark_ops/benchmark_rmsnorm.py
examples/inference/benchmark_ops/benchmark_rmsnorm.py
import torch from colossalai.kernel.kernel_loader import InferenceOpsLoader from colossalai.kernel.triton import rms_layernorm try: import triton # noqa except ImportError: print("please install triton from https://github.com/openai/triton") inference_ops = InferenceOpsLoader().load() # Triton benchmark plot attributions configs = [ triton.testing.Benchmark( x_names=["SEQUENCE_TOTAL"], x_vals=[i for i in range(128, 1025, 128)], line_arg="provider", line_vals=[ "vllm_rms_layernorm", "triton_rms_layernorm", "cuda_rms_layernorm", "vllm_rms_layernorm_with_residual", "triton_rms_layernorm_with_residual", "cuda_rms_layernorm_with_residual", ], line_names=[ "vllm_rms_layernorm", "triton_rms_layernorm", "cuda_rms_layernorm", "vllm_rms_layernorm_with_residual", "triton_rms_layernorm_with_residual", "cuda_rms_layernorm_with_residual", ], styles=[("red", "-"), ("blue", "-"), ("yellow", "-"), ("red", "--"), ("blue", "--"), ("yellow", "--")], ylabel="ms", plot_name=f"RMSNorm benchmarking results", args={"HIDDEN_SIZE": 5120}, ) ] @triton.testing.perf_report(configs) def benchmark_rms_layernorm( provider: str, SEQUENCE_TOTAL: int, HIDDEN_SIZE: int, ): try: from vllm.model_executor.layers.layernorm import RMSNorm except ImportError: raise ImportError("Please install vllm from https://github.com/vllm-project/vllm") warmup = 10 rep = 1000 dtype = torch.float16 eps = 1e-5 x_shape = (SEQUENCE_TOTAL, HIDDEN_SIZE) w_shape = (x_shape[-1],) residual = torch.rand(x_shape, dtype=dtype, device="cuda") weight = torch.ones(w_shape, dtype=dtype, device="cuda") vllm_norm = RMSNorm(hidden_size=HIDDEN_SIZE, eps=eps).to(dtype=dtype, device="cuda") x = -2.3 + 0.5 * torch.randn(x_shape, dtype=dtype, device="cuda") if provider == "vllm_rms_layernorm": fn = lambda: vllm_norm(x) elif provider == "triton_rms_layernorm": fn = lambda: rms_layernorm(x, weight, eps=eps) elif provider == "cuda_rms_layernorm": out = torch.empty_like(x) fn = lambda: inference_ops.rms_layernorm(out, x, weight, eps) elif provider == "vllm_rms_layernorm_with_residual": fn = lambda: vllm_norm(x, residual=residual) elif provider == "triton_rms_layernorm_with_residual": fn = lambda: rms_layernorm(x, weight, eps=eps, residual=residual) elif provider == "cuda_rms_layernorm_with_residual": fn = lambda: inference_ops.fused_add_rms_layernorm(x, residual, weight, eps) else: raise ValueError("Undefined provider.") ms = triton.testing.do_bench(fn, warmup=warmup, rep=rep) return ms if __name__ == "__main__": benchmark_rms_layernorm.run(save_path=".", print_data=True)
python
Apache-2.0
b1915d2889543949eb5b610241f1515c73df5059
2026-01-04T14:40:19.002665Z
false