python_code
stringlengths
0
108k
#!/usr/bin/env python # -*- encoding: utf-8 -*- import time import torch from colossalai.constants import INPUT_GROUP_3D, OUTPUT_GROUP_3D, WEIGHT_GROUP_3D from colossalai.core import global_context from colossalai.logging import get_dist_logger from colossalai.nn import (Classifier3D, CrossEntropyLoss3D, Embedding3D,...
#!/usr/bin/env python # -*- encoding: utf-8 -*- train_data = dict( dataset=dict( type='CIFAR10Dataset', root='/path/to/data', download=True, transform_pipeline=[ dict(type='RandomResizedCrop', size=224), dict(type='RandomHorizontalFlip'), dict(typ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from pathlib import Path import pytest from colossalai.context.config import Config from colossalai.builder import build_ophooks @pytest.mark.cpu def test_load_config(): filename = Path(__file__).parent.joinpath('sample_config.py') config = Config.from_file(f...
import torch import colossalai import torch.multiprocessing as mp from colossalai.amp import convert_to_naive_amp, convert_to_apex_amp from tests.components_to_test.registry import non_distributed_component_funcs from colossalai.testing import assert_close_loose, rerun_if_address_is_in_use from colossalai.utils import ...
import os from functools import partial from pathlib import Path import colossalai import pytest import torch import torch.multiprocessing as mp from colossalai.amp.amp_type import AMP_TYPE from colossalai.builder import build_pipeline_model from colossalai.engine.schedule import PipelineSchedule from colossalai.loggi...
from functools import partial import colossalai import pytest import torch.multiprocessing as mp from colossalai.amp import AMP_TYPE from colossalai.core import global_context as gpc from colossalai.utils import free_port from tests.components_to_test.registry import non_distributed_component_funcs from colossalai.tes...
import pytest from colossalai.engine.paramhooks import BaseParamHookMgr from torch import nn import torch import torch.nn.functional as F import copy class SubNet(nn.Module): def __init__(self, out_features) -> None: super().__init__() self.bias = nn.Parameter(torch.zeros(out_features)) def fo...
#!/usr/bin/env python class Registry: def __init__(self): self._registry = dict() def register(self, name): assert name not in self._registry def _regsiter(callable_): self._registry[name] = callable_ return _regsiter def get_callable(self, name: str): ...
import torch import torch.nn as nn import torch.nn.functional as F from colossalai.nn import CheckpointModule from .utils import DummyDataGenerator from .registry import non_distributed_component_funcs class SubNet(nn.Module): def __init__(self, out_features) -> None: super().__init__() self.bias...
from . import repeated_computed_layer, resnet, nested_model, bert, no_leaf_module
from torchvision.models import resnet18 from .registry import non_distributed_component_funcs from pathlib import Path import os import torch from torchvision.transforms import transforms from torchvision.datasets import CIFAR10 from colossalai.utils import get_dataloader def get_cifar10_dataloader(train): # buil...
import torch import torch.nn as nn import torch.nn.functional as F from colossalai.nn import CheckpointModule from .utils.dummy_data_generator import DummyDataGenerator from .registry import non_distributed_component_funcs class NoLeafModule(CheckpointModule): """ In this no-leaf module, it has subordinate nn...
#!/usr/bin/env python import torch import torch.nn as nn from colossalai.nn import CheckpointModule from .utils.dummy_data_generator import DummyDataGenerator from .registry import non_distributed_component_funcs class NetWithRepeatedlyComputedLayers(CheckpointModule): """ This model is to test with layers w...
import torch import transformers from packaging import version from torch.utils.data import SequentialSampler from transformers import BertConfig, BertForSequenceClassification from .registry import non_distributed_component_funcs def get_bert_data_loader( batch_size, total_samples, sequence_...
from abc import ABC, abstractmethod class DummyDataGenerator(ABC): def __init__(self, length=10): self.length = length @abstractmethod def generate(self): pass def __iter__(self): self.step = 0 return self def __next__(self): if self.step < self.length: ...
from .dummy_data_generator import DummyDataGenerator
import torch import torch.nn as nn from torch.optim.adam import Adam from torch.optim import AdamW from colossalai.nn.optimizer.fused_adam import FusedAdam from colossalai.testing import parameterize class FC(nn.Module): def __init__(self) -> None: super().__init__() self.fc = nn.Sequential(nn.Li...
import math import torch from colossalai.testing import parameterize def torch_adam_update( step, lr, beta1, beta2, eps, weight_decay, param, grad, exp_avg, exp_avg_sq, use_adamw, ): bias_correction1 = 1 - beta1**step bias_correction2 = 1 - beta2**step if weig...
import torch import torch.nn as nn from torch.optim.adam import Adam from torch.optim import AdamW from colossalai.nn.optimizer.hybrid_adam import HybridAdam from colossalai.testing import parameterize RE = 1024 @parameterize('adamw', [False, True]) @parameterize('device', ['cpu', 'cuda:0']) @parameterize('p_dtype'...
from numpy import dtype import torch import torch.nn as nn import math from colossalai.testing import parameterize from colossalai.utils import multi_tensor_applier def torch_adam_update( step, lr, beta1, beta2, eps, weight_decay, param, grad, exp_avg, exp_avg_sq, use_adam...
from functools import partial import colossalai import pytest import torch import torch.multiprocessing as mp from colossalai.amp.amp_type import AMP_TYPE from colossalai.logging import get_dist_logger from colossalai.trainer import Trainer from colossalai.utils import MultiTimer, free_port from tests.components_to_te...
import os from functools import partial from pathlib import Path import colossalai import pytest import torch import torch.multiprocessing as mp import torch.nn as nn from colossalai.context.parallel_mode import ParallelMode from colossalai.core import global_context as gpc from colossalai.engine.schedule import Pipel...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from functools import partial import pytest import torch import torch.distributed as dist import torch.multiprocessing as mp from colossalai.communication import (recv_backward, recv_forward, recv_tensor_meta, send_backward, send_ba...
# referenced from Megatron and used to testify communication import os import os.path as osp from functools import partial from pathlib import Path import colossalai import pytest import torch import torch.multiprocessing as mp from colossalai.builder import build_pipeline_model_from_cfg from colossalai.core import g...
import os.path as osp import pytest import torch import torch.multiprocessing as mp from colossalai.builder.pipeline import build_pipeline_model_from_cfg from colossalai.core import global_context from colossalai.initialize import launch from colossalai.logging import get_dist_logger from functools import partial fro...
import os import model from pathlib import Path BATCH_SIZE = 128 IMG_SIZE = 224 DIM = 768 NUM_CLASSES = 10 NUM_ATTN_HEADS = 12 NUM_MICRO_BATCHES = 2 # resnet 18 model = dict(type='VanillaResNet', block_type='ResNetBasicBlock', layers=[2, 2, 2, 2], num_cls=10) parallel = dict( ...
from .layers import * from .resnet import VanillaResNet
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import List, Optional import torch import torch.nn as nn from torch import Tensor from colossalai.registry import LAYERS from colossalai.registry import MODELS from colossalai.nn.model import ModelFromConfig @MODELS.register_module class VanillaResNet(Mod...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Optional, Callable import torch.nn as nn from torch import Tensor from colossalai.registry import LAYERS from .conv import conv3x3 @LAYERS.register_module class ResNetBasicBlock(nn.Module): """Basic ResNet block """ expansion: int = 1 ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.nn as nn from colossalai.registry import LAYERS from .conv import conv1x1 @LAYERS.register_module class ResLayer(nn.Module): def __init__(self, block_type: str, norm_layer_type: str, inplanes: int, ...
from .basic_block import ResNetBasicBlock from .bottleneck import ResNetBottleneck from .reslayer import ResLayer
#!/usr/bin/env python # -*- encoding: utf-8 -*- from typing import Optional, Callable import torch.nn as nn from torch import Tensor from colossalai.registry import LAYERS from .conv import conv3x3, conv1x1 @LAYERS.register_module class ResNetBottleneck(nn.Module): # Bottleneck in torchvision places the stride...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.nn as nn def conv3x3(in_planes: int, out_planes: int, stride: int = 1, groups: int = 1, dilation: int = 1) -> nn.Conv2d: """3x3 convolution with padding""" return nn.Conv2d(in_planes, out_planes, kernel_size=3, stride=stride, p...
from functools import partial import pytest import torch import torch.distributed as dist import torch.multiprocessing as mp from colossalai.communication import all_gather, all_reduce, reduce_scatter from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.initiali...
from functools import partial import colossalai import pytest import torch import torch.multiprocessing as mp from colossalai.amp import convert_to_apex_amp from colossalai.nn.optimizer import CPUAdam from colossalai.testing import parameterize, rerun_if_address_is_in_use from colossalai.utils import free_port from co...
from functools import partial import colossalai import pytest import torch import torch.multiprocessing as mp from colossalai.testing import parameterize, rerun_if_address_is_in_use from colossalai.utils import free_port from colossalai.zero.init_ctx import ZeroInitContext from colossalai.zero.shard_utils import (Buck...
from functools import partial import pytest import torch import torch.nn as nn import torch.multiprocessing as mp import colossalai from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.utils import free_port, get_current_device from colossalai.nn.layer.moe import...
from functools import partial import pytest import torch.nn as nn import torch.multiprocessing as mp import torch.distributed as dist import colossalai from colossalai.utils import free_port, get_current_device from colossalai.nn.layer.moe import Experts from colossalai.context.moe_context import MOE_CONTEXT from colos...
from functools import partial import colossalai import pytest import torch import torch.multiprocessing as mp import torch.nn as nn from colossalai.nn import CheckpointModule from colossalai.logging import get_dist_logger from colossalai.testing import parameterize from colossalai.utils import free_port from colossala...
from functools import partial import pytest import torch import torch.nn as nn import torch.multiprocessing as mp import torch.distributed as dist import colossalai from colossalai.utils import free_port, get_current_device from colossalai.nn.layer.moe import Top1Router, UniformNoiseGenerator, MoeLayer, Experts from co...
# Configuration file for the Sphinx documentation builder. # # This file only contains a selection of the most common options. For a full # list see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html # -- Path setup -------------------------------------------------------------- import ...
import torch import torch.nn as nn from colossalai.nn.layer import WrappedDropPath as DropPath class TransformerLayer(nn.Module): """Transformer layer builder. """ def __init__(self, att: nn.Module, ffn: nn.Module, norm1: nn.Module, norm2...
from .gpt import *
import math from typing import Callable import torch from colossalai import nn as col_nn from colossalai.builder.pipeline import partition_uniform from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.logging import get_dist_logger from colossalai.nn.layer.utils ...
from .vit import *
import math from typing import Callable import torch from colossalai import nn as col_nn from colossalai.nn.layer.utils import CheckpointModule from colossalai.registry import LAYERS, MODELS from torch import dtype, nn __all__ = [ 'VisionTransformer', 'vit_lite_depth7_patch4_32', 'vit_tiny_patch4_32', ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch from colossalai.registry import MODELS from colossalai.nn.model.model_from_config import ModelFromConfig @MODELS.register_module class VisionTransformerFromConfig(ModelFromConfig): """Vision Transformer from `"An Image is Worth 16x16 Words: Trans...
import math import torch import torch.nn as nn from colossalai.context import ParallelMode from colossalai.nn.layer import VanillaPatchEmbedding, VanillaClassifier, \ WrappedDropout as Dropout, WrappedDropPath as DropPath from colossalai.nn.layer.moe import build_ffn_experts, MoeLayer, Top2Router, NormalNoiseGenera...
from colossalai.context import ParallelMode from colossalai.nn.layer import WrappedDropout as Dropout def moe_sa_args(d_model: int, n_heads: int, d_kv: int, attention_drop: float = 0, drop_rate: float = 0, bias: bool = True): """This ...
from .models import Widenet, ViTMoE from .gpt import MOEGPT, prmoe_4b, prmoe_31b, prmoe_51b
from typing import Callable, List from torch import dtype, nn from colossalai import nn as col_nn from colossalai.registry import LAYERS, MODELS from colossalai.nn.layer import MoeModule from colossalai.context import MOE_CONTEXT from colossalai.logging import get_dist_logger from colossalai.nn.layer.utils import Check...
from typing import Optional class TensorParallelEnv(object): _instance = None def __new__(cls, *args, **kwargs): if cls._instance is None: cls._instance = object.__new__(cls, *args, **kwargs) return cls._instance def __init__(self, *args, **kwargs): self.load(*args, *...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import argparse import os import pprint from pathlib import Path from typing import Callable, Dict, Iterable, List, Optional, Tuple, Union import torch import torch.nn as nn from torch.nn.modules.loss import _Loss from torch.nn.parallel import DistributedDataParallel as...
#!/usr/bin/env python # -*- encoding: utf-8 -*- ALLOWED_MODES = [None, '1d', '2d', '2.5d', '3d', 'sequence'] TENSOR_PARALLEL_MODE = 'tensor_parallel_mode' # intializer INITIALIZER_MAPPING = { 'data': 'Initializer_Data', 'tensor': 'Initializer_Tensor', 'pipeline': 'Initializer_Pipeline', 'embedding': '...
from .initialize import (initialize, launch, launch_from_openmpi, launch_from_slurm, launch_from_torch, get_default_parser) __version__ = '0.0.1'
#!/usr/bin/env python # -*- encoding: utf-8 -*- from colossalai.context.parallel_context import global_context
#!/usr/bin/env python # -*- encoding: utf-8 -*- import inspect import sys from importlib.machinery import SourceFileLoader from pathlib import Path from colossalai.logging import get_dist_logger class Config(dict): """This is a wrapper class for dict objects so that values of which can be accessed as attribu...
from .config import Config, ConfigException from .parallel_context import ParallelContext from .parallel_mode import ParallelMode from .moe_context import MOE_CONTEXT from .process_group_initializer import * from .random import *
#!/usr/bin/env python # -*- encoding: utf-8 -*- import random import socket from collections import Counter from typing import Union import numpy as np import torch import torch.distributed as dist from colossalai.constants import ALLOWED_MODES, INITIALIZER_MAPPING from colossalai.context.config import Config from co...
import torch import torch.distributed as dist from colossalai.context.parallel_mode import ParallelMode from colossalai.context.singleton_meta import SingletonMeta from typing import Tuple def _check_sanity(): from colossalai.core import global_context as gpc if gpc.tensor_parallel_size > 1 or gpc.pipeline_...
class SingletonMeta(type): """ The Singleton class can be implemented in different ways in Python. Some possible methods include: base class, decorator, metaclass. We will use the metaclass because it is best suited for this purpose. """ _instances = {} def __call__(cls, *args, **kwargs): ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from enum import Enum # parallel modes class ParallelMode(Enum): """This is an enumeration class containing all possible parallel modes. """ GLOBAL = 'global' # common parallel DATA = 'data' # model parallel - containing tensor and pipeline p...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from torch import distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode @DIST_GROUP_INITIALIZER.register_module class Initializer_Pipeline(Proc...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math import torch.distributed as dist from colossalai.global_variables import tensor_parallel_env as env from colossalai.registry import DIST_GROUP_INITIALIZER from ..parallel_mode import ParallelMode from .process_group_initializer import ProcessGroupInitialize...
import math import torch.distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode from colossalai.global_variables import tensor_parallel_env as env def _check_summa_env_var(summa_dim): # ch...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.distributed as dist from colossalai.global_variables import tensor_parallel_env as env from colossalai.registry import DIST_GROUP_INITIALIZER from ..parallel_mode import ParallelMode from .process_group_initializer import ProcessGroupInitializer @DIST_GRO...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode @DIST_GROUP_INITIALIZER.register_module class Initializer_Tensor(ProcessGrou...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode @DIST_GROUP_INITIALIZER.register_module class Initializer_Model(ProcessGroupI...
from .initializer_1d import Initializer_1D from .initializer_2d import Initializer_2D from .initializer_2p5d import Initializer_2p5D from .initializer_3d import Initializer_3D from .initializer_data import Initializer_Data from .initializer_pipeline import Initializer_Pipeline from .initializer_sequence import Initiali...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from abc import ABC, abstractmethod from colossalai.context import Config class ProcessGroupInitializer(ABC): """An object, knowing the parallelism configuration, that initializes parallel groups. Args: rank (int): The rank of current process. ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math import torch.distributed as dist from colossalai.context import Config from colossalai.global_variables import tensor_parallel_env as env from colossalai.registry import DIST_GROUP_INITIALIZER from ..parallel_mode import ParallelMode from .process_group_ini...
#!/usr/bin/env python # -*- encoding: utf-8 -*- from torch import distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode @DIST_GROUP_INITIALIZER.register_module class Initializer_Data(ProcessG...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch.distributed as dist from colossalai.registry import DIST_GROUP_INITIALIZER from .initializer_tensor import Initializer_Tensor from .process_group_initializer import ProcessGroupInitializer from ..parallel_mode import ParallelMode @DIST_GROUP_INITIALIZER.re...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch from torch import Tensor from colossalai.context.parallel_mode import ParallelMode class SeedManager: """This class is a manager of all random seeds involved in the system. Note: The parallel_mode should be concluded in ``ParallelMode``. ...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import functools from contextlib import contextmanager import torch.cuda from torch import Tensor from .seed_manager import SeedManager from ..parallel_mode import ParallelMode _SEED_MANAGER = SeedManager() def get_seeds(): """Returns the seeds of the seed manag...
from ._helper import (seed, set_mode, with_seed, add_seed, get_seeds, get_states, get_current_mode, set_seed_states, sync_states, moe_set_seed, reset_seeds) __all__ = [ 'seed', 'set_mode', 'with_seed', 'add_seed', 'get_seeds', 'get_states', 'get_current_mode', 'set_seed_states', 'sync_sta...
from .stateful_tensor_mgr import StatefulTensorMgr from .tensor_placement_policy import TensorPlacementPolicyFactory __all__ = ['StatefulTensorMgr', 'TensorPlacementPolicyFactory']
from abc import ABC, abstractmethod from typing import List, Optional import torch from colossalai.utils import get_current_device from colossalai.zero.sharded_param.tensor_utils import colo_model_data_tensor_move_inline, colo_tensor_mem_usage from colossalai.utils.memory import colo_device_memory_capacity from colossa...
import functools import torch import types from colossalai.utils.cuda import get_current_device from colossalai.zero.sharded_param.sharded_param import ShardedParamV2 from colossalai.zero.sharded_param.tensorful_state import StatefulTensor, TensorState from colossalai.zero.sharded_param.tensor_utils import colo_model_d...
from .layer import * from .loss import * from .lr_scheduler import * from .metric import * from .model import * from .optimizer import *
import math import warnings from torch import Tensor import torch.nn as nn def zeros_(): """Return the initializer filling the input Tensor with the scalar zeros""" def initializer(tensor: Tensor, fan_in: int = None, fan_out: int = None): return nn.init.zeros_(tensor) return initializer def on...
# modified from https://github.com/NVIDIA/apex/blob/master/apex/optimizers/fused_adam.py import torch from colossalai.registry import OPTIMIZERS from colossalai.utils import multi_tensor_applier @OPTIMIZERS.register_module class FusedAdam(torch.optim.Optimizer): """Implements Adam algorithm. Currently GPU-o...
""" Adapted from the pytorch-lamb library at https://github.com/cybertronai/pytorch-lamb """ import torch from torch.optim import Optimizer from colossalai.registry import OPTIMIZERS @OPTIMIZERS.register_module class Lamb(Optimizer): r"""Implements Lamb algorithm. It has been proposed in `Large Batch Optimi...
"""Adapted from https://github.com/NUS-HPC-AI-Lab/LARS-ImageNet-PyTorch/blob/main/lars.py""" from typing import Iterable import torch from torch.optim import Optimizer from colossalai.registry import OPTIMIZERS @OPTIMIZERS.register_module class Lars(Optimizer): r"""Implements the LARS optimizer from `"Large ba...
from .utils import CPU_ADAM_CNT from .colossalai_optimizer import ColossalaiOptimizer from .fused_adam import FusedAdam from .fused_lamb import FusedLAMB from .fused_sgd import FusedSGD from .lamb import Lamb from .lars import Lars from .cpu_adam import CPUAdam from .hybrid_adam import HybridAdam __all__ = ['Colossala...
import math import torch from colossalai.registry import OPTIMIZERS from colossalai.nn.optimizer import CPU_ADAM_CNT @OPTIMIZERS.register_module class CPUAdam(torch.optim.Optimizer): """Implements Adam algorithm. Supports parameters updating on both GPU and CPU, depanding on the device of paramters. But...
class CpuAdamCounter(object): """Used to record the total number of CPU Adam. We must use it to avoid hybrid cpu adam and cpu adam using the same id. """ def __init__(self): self.number = 0 def __call__(self): self.number += 1 return self.number - 1 CPU_ADAM_CNT = CpuAdam...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import torch import torch.nn as nn from torch import Tensor from torch.optim import Optimizer from colossalai.utils import clip_grad_norm_fp32 class ColossalaiOptimizer(Optimizer): def __init__(self, optim: Optimizer): self.optim = optim @property ...
# modified from https://github.com/NVIDIA/apex/blob/master/apex/optimizers/fused_lamb.py import torch from colossalai.registry import OPTIMIZERS from colossalai.utils import multi_tensor_applier @OPTIMIZERS.register_module class FusedLAMB(torch.optim.Optimizer): """Implements LAMB algorithm. Currently GPU-o...
import torch from colossalai.utils import multi_tensor_applier from colossalai.registry import OPTIMIZERS from colossalai.nn.optimizer import CPU_ADAM_CNT @OPTIMIZERS.register_module class HybridAdam(torch.optim.Optimizer): """Implements Adam algorithm. Supports parameters updating on both GPU and CPU, depa...
# modified from https://github.com/NVIDIA/apex/blob/master/apex/optimizers/fused_sgd.py import torch from torch.optim.optimizer import Optimizer, required from colossalai.registry import OPTIMIZERS from colossalai.utils import multi_tensor_applier @OPTIMIZERS.register_module class FusedSGD(Optimizer): r"""Implem...
import torch.nn as nn from colossalai.registry import LOSSES from torch.nn.modules.loss import _Loss from colossalai.context.moe_context import MOE_CONTEXT @LOSSES.register_module class MoeCrossEntropyLoss(_Loss): r"""torch.nn.CrossEntropyLoss added with auxiliary loss. Args: input (:class:`torch.ten...
from colossalai.global_variables import tensor_parallel_env as env from colossalai.nn.layer.utils import get_tensor_parallel_mode from torch import nn from torch.nn.modules.loss import * from torch.nn.modules.loss import _Loss from .loss_1d import VocabParallelCrossEntropyLoss1D from .loss_2d import CrossEntropyLoss2D...
import torch from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.registry import LOSSES from torch.cuda.amp import custom_bwd, custom_fwd from torch.nn.modules.loss import _Loss class _VocabParallelCrossEntropy1D(torch.autograd.Function): @staticmethod ...
import torch import torch.distributed as dist from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.nn.layer.parallel_2d import reduce_by_batch_2d, split_batch_2d from colossalai.nn.layer.parallel_2d._utils import assert_summa_initialization from colossalai.regist...
import torch import torch.distributed as dist from colossalai.constants import INPUT_GROUP_3D, WEIGHT_GROUP_3D, OUTPUT_GROUP_3D from colossalai.core import global_context as gpc from colossalai.nn.layer.parallel_3d import reduce_by_batch_3d, split_tensor_3d from colossalai.nn.layer.parallel_3d._utils import get_paralle...
import torch import torch.distributed as dist from colossalai.context import ParallelMode from colossalai.core import global_context as gpc from colossalai.nn.layer.parallel_2p5d import reduce_by_batch_2p5d, split_batch_2p5d from colossalai.nn.layer.parallel_2p5d._utils import assert_tesseract_initialization from colos...
from torch.optim.lr_scheduler import CosineAnnealingLR as _CosineAnnealingLR from colossalai.registry import LR_SCHEDULERS from .delayed import DelayerScheduler, WarmupDelayerScheduler, WarmupScheduler @LR_SCHEDULERS.register_module class CosineAnnealingLR(_CosineAnnealingLR): r"""Set the learning rate of each p...
from torch.optim.lr_scheduler import _LRScheduler from colossalai.registry import LR_SCHEDULERS @LR_SCHEDULERS.register_module class LinearWarmupLR(_LRScheduler): """Linearly warmup learning rate and then linearly decay. Args: optimizer (:class:`torch.optim.Optimizer`): Wrapped optimizer. to...