python_code
stringlengths
0
456k
from inspect import isfunction import math import torch import torch.nn.functional as F from torch import nn, einsum from einops import rearrange, repeat from typing import Optional, Any from ldm.modules.diffusionmodules.util import checkpoint try: import xformers import xformers.ops XFORMERS_IS_AVAILBLE...
import torch from torch import nn class LitEma(nn.Module): def __init__(self, model, decay=0.9999, use_num_upates=True): super().__init__() if decay < 0.0 or decay > 1.0: raise ValueError('Decay must be between 0 and 1') self.m_name2s_name = {} self.register_buffer('de...
# based on https://github.com/isl-org/MiDaS import cv2 import torch import torch.nn as nn from torchvision.transforms import Compose from ldm.modules.midas.midas.dpt_depth import DPTDepthModel from ldm.modules.midas.midas.midas_net import MidasNet from ldm.modules.midas.midas.midas_net_custom import MidasNet_small fr...
"""Utils for monoDepth.""" import sys import re import numpy as np import cv2 import torch def read_pfm(path): """Read pfm file. Args: path (str): path to file Returns: tuple: (data, scale) """ with open(path, "rb") as file: color = None width = None heig...
"""MidashNet: Network for monocular depth estimation trained by mixing several datasets. This file contains code that is adapted from https://github.com/thomasjpfan/pytorch_refinenet/blob/master/pytorch_refinenet/refinenet/refinenet_4cascade.py """ import torch import torch.nn as nn from .base_model import BaseModel f...
import numpy as np import cv2 import math def apply_min_size(sample, size, image_interpolation_method=cv2.INTER_AREA): """Rezise the sample to ensure the given size. Keeps aspect ratio. Args: sample (dict): sample size (tuple): image size Returns: tuple: new size """ shap...
"""MidashNet: Network for monocular depth estimation trained by mixing several datasets. This file contains code that is adapted from https://github.com/thomasjpfan/pytorch_refinenet/blob/master/pytorch_refinenet/refinenet/refinenet_4cascade.py """ import torch import torch.nn as nn from .base_model import BaseModel f...
import torch import torch.nn as nn import torch.nn.functional as F from .base_model import BaseModel from .blocks import ( FeatureFusionBlock, FeatureFusionBlock_custom, Interpolate, _make_encoder, forward_vit, ) def _make_fusion_block(features, use_bn): return FeatureFusionBlock_custom( ...
import torch class BaseModel(torch.nn.Module): def load(self, path): """Load model from file. Args: path (str): file path """ parameters = torch.load(path, map_location=torch.device('cpu')) if "optimizer" in parameters: parameters = parameters["mod...
import torch import torch.nn as nn import timm import types import math import torch.nn.functional as F class Slice(nn.Module): def __init__(self, start_index=1): super(Slice, self).__init__() self.start_index = start_index def forward(self, x): return x[:, self.start_index :] class...
import torch import torch.nn as nn from .vit import ( _make_pretrained_vitb_rn50_384, _make_pretrained_vitl16_384, _make_pretrained_vitb16_384, forward_vit, ) def _make_encoder(backbone, features, use_pretrained, groups=1, expand=False, exportable=True, hooks=None, use_vit_only=False, use_readout="ign...
import torch import torch.nn as nn from torch.utils.checkpoint import checkpoint from transformers import T5Tokenizer, T5EncoderModel, CLIPTokenizer, CLIPTextModel import open_clip from ldm.util import default, count_params class AbstractEncoder(nn.Module): def __init__(self): super().__init__() de...
# -*- coding: utf-8 -*- """ # -------------------------------------------- # Super-Resolution # -------------------------------------------- # # Kai Zhang (cskaizhang@gmail.com) # https://github.com/cszn # From 2019/03--2021/08 # -------------------------------------------- """ import numpy as np import cv2 import tor...
from ldm.modules.image_degradation.bsrgan import degradation_bsrgan_variant as degradation_fn_bsr from ldm.modules.image_degradation.bsrgan_light import degradation_bsrgan_variant as degradation_fn_bsr_light
import os import math import random import numpy as np import torch import cv2 from torchvision.utils import make_grid from datetime import datetime #import matplotlib.pyplot as plt # TODO: check with Dominik, also bsrgan.py vs bsrgan_light.py os.environ["KMP_DUPLICATE_LIB_OK"]="TRUE" ''' # ----------------------...
# -*- coding: utf-8 -*- import numpy as np import cv2 import torch from functools import partial import random from scipy import ndimage import scipy import scipy.stats as ss from scipy.interpolate import interp2d from scipy.linalg import orth import albumentations import ldm.modules.image_degradation.utils_image as ...
# adopted from # https://github.com/openai/improved-diffusion/blob/main/improved_diffusion/gaussian_diffusion.py # and # https://github.com/lucidrains/denoising-diffusion-pytorch/blob/7706bdfc6f527f58d33f84b7b522e61e6e3164b3/denoising_diffusion_pytorch/denoising_diffusion_pytorch.py # and # https://github.com/openai/gu...
import torch import torch.nn as nn import numpy as np from functools import partial from ldm.modules.diffusionmodules.util import extract_into_tensor, make_beta_schedule from ldm.util import default class AbstractLowScaleModel(nn.Module): # for concatenating a downsampled image to the latent representation d...
# pytorch_diffusion + derived encoder decoder import math from typing import Any, Optional import numpy as np import torch import torch.nn as nn from einops import rearrange try: from lightning.pytorch.utilities import rank_zero_info except: from pytorch_lightning.utilities import rank_zero_info from ldm.mod...
from abc import abstractmethod import math import numpy as np import torch as th import torch.nn as nn import torch.nn.functional as F from ldm.modules.diffusionmodules.util import ( checkpoint, conv_nd, linear, avg_pool_nd, zero_module, normalization, timestep_embedding, ) from ldm.module...
import torch import numpy as np class AbstractDistribution: def sample(self): raise NotImplementedError() def mode(self): raise NotImplementedError() class DiracDistribution(AbstractDistribution): def __init__(self, value): self.value = value def sample(self): retur...
import os import numpy as np import PIL from PIL import Image from torch.utils.data import Dataset from torchvision import transforms class LSUNBase(Dataset): def __init__(self, txt_file, data_root, size=None, interpolation="bicubic", ...
import os, yaml, pickle, shutil, tarfile, glob import cv2 import albumentations import PIL import numpy as np import torchvision.transforms.functional as TF from omegaconf import OmegaConf from functools import partial from PIL import Image from tqdm import tqdm from torch.utils.data import Dataset, Subset import tami...
from typing import Dict import numpy as np from omegaconf import DictConfig, ListConfig import torch from torch.utils.data import Dataset from pathlib import Path import json from PIL import Image from torchvision import transforms from einops import rearrange from ldm.util import instantiate_from_config from datasets ...
import math import os from abc import abstractmethod import cv2 import numpy as np import torch from torch.utils.data import ChainDataset, ConcatDataset, Dataset, IterableDataset class Txt2ImgIterableBaseDataset(IterableDataset): ''' Define an interface to make the IterableDatasets for text2img data chainabl...
from typing import Dict import numpy as np from omegaconf import DictConfig, ListConfig import torch from torch.utils.data import Dataset from pathlib import Path import json from PIL import Image from torchvision import transforms from einops import rearrange from ldm.util import instantiate_from_config from datasets ...
from abc import ABC, abstractmethod import torch import torch.nn as nn from transformers import ViTConfig, ViTForImageClassification from colossalai.utils.cuda import get_current_device class DummyDataGenerator(ABC): def __init__(self, length=10): self.length = length @abstractmethod def gener...
import os import torch import torch.distributed as dist import torch.nn as nn import torch.nn.functional as F from timm.models.vision_transformer import _create_vision_transformer from titans.dataloader.imagenet import build_dali_imagenet from tqdm import tqdm from vit import DummyDataLoader import colossalai from co...
import os import random from functools import partial import numpy as np import pytest import torch import torch.multiprocessing as mp from torch.nn.parallel import DistributedDataParallel as DDP from vit import get_training_components import colossalai from colossalai.context import ParallelMode from colossalai.cont...
from colossalai.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 8 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 3 WARMUP_EPOCHS = 1 # model config IMG_SIZE = 224 PATCH_SIZE = 16 HIDDEN_SIZE = 32 DEPTH = 2 NUM_HEADS = 4 MLP_RATI...
from colossalai.amp import AMP_TYPE # hyperparameters # BATCH_SIZE is as per GPU # global batch size = BATCH_SIZE x data parallel size BATCH_SIZE = 256 LEARNING_RATE = 3e-3 WEIGHT_DECAY = 0.3 NUM_EPOCHS = 300 WARMUP_EPOCHS = 32 # model config IMG_SIZE = 224 PATCH_SIZE = 16 HIDDEN_SIZE = 384 DEPTH = 12 NUM_HEADS = 6 M...
import argparse import hashlib import itertools import math import os from pathlib import Path from typing import Optional import torch import torch.nn.functional as F import torch.utils.checkpoint from accelerate import Accelerator from accelerate.logging import get_logger from accelerate.utils import set_seed from d...
import argparse import hashlib import itertools import math import os import random from pathlib import Path from typing import Optional import numpy as np import torch import torch.nn.functional as F import torch.utils.checkpoint from accelerate import Accelerator from accelerate.logging import get_logger from accele...
''' torchrun --standalone --nproc_per_node=1 debug.py ''' from diffusers import AutoencoderKL import colossalai from colossalai.utils.model.colo_init_context import ColoInitContext, post_process_colo_init_ctx path = "/data/scratch/diffuser/stable-diffusion-v1-4" colossalai.launch_from_torch(config={}) with ColoInit...
from diffusers import StableDiffusionPipeline, DiffusionPipeline import torch model_id = <Your Model Path> print(f"Loading model... from{model_id}") pipe = DiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float16).to("cuda") prompt = "A photo of an apple." image = pipe(prompt, num_inference_steps=50, gu...
import argparse import hashlib import math import os from pathlib import Path from typing import Optional import torch import torch.nn.functional as F import torch.utils.checkpoint from diffusers import AutoencoderKL, DDPMScheduler, DiffusionPipeline, UNet2DConditionModel from diffusers.optimization import get_schedul...
import contextlib import os import torch import torch.nn as nn from dataset.webtext import WebtextDataset from titans.model.gpt import GPTLMLoss import colossalai import colossalai.utils as utils from colossalai.context.parallel_mode import ParallelMode from colossalai.core import global_context as gpc from colossala...
import json import os from typing import Optional import torch from torch.utils.data import Dataset from transformers import GPT2Tokenizer from colossalai.registry import DATASETS @DATASETS.register_module class WebtextDataset(Dataset): def __init__(self, path: Optional[str] = None, seq_len=1024) -> None: ...
from model import GPT3_pipeline_hybrid from colossalai.nn.optimizer import HybridAdam from colossalai.zero.shard_utils import TensorShardStrategy BATCH_SIZE = 192 NUM_EPOCHS = 60 SEQ_LEN = 2048 NUM_MICRO_BATCHES = 192 HIDDEN_SIZE = 12288 TENSOR_SHAPE = (BATCH_SIZE // NUM_MICRO_BATCHES, SEQ_LEN, HIDDEN_SIZE) # if you...
from model import GPT2_small_pipeline_hybrid from colossalai.nn.optimizer import HybridAdam from colossalai.zero.shard_utils import TensorShardStrategy BATCH_SIZE = 8 NUM_EPOCHS = 10 SEQ_LEN = 1024 NUM_MICRO_BATCHES = 4 HIDDEN_SIZE = 768 TENSOR_SHAPE = (BATCH_SIZE // NUM_MICRO_BATCHES, SEQ_LEN, HIDDEN_SIZE) # if you...
import torch import torch.nn.init as init from torch import Tensor from torch import distributed as dist from torch import nn as nn from torch.nn import functional as F from torch.nn.parameter import Parameter from colossalai.context import ParallelMode, seed from colossalai.core import global_context as gpc from colo...
from .embed import vocab_parallel_cross_entropy from .gpt1d import * from .pipeline_gpt1d import *
import inspect # import model_zoo.gpt.gpt as col_gpt import titans.model.gpt.gpt as col_gpt import torch import torch.nn as nn from colossalai import kernel from colossalai import nn as col_nn from colossalai.context.parallel_mode import ParallelMode from colossalai.core import global_context as gpc from colossalai.l...
#!/usr/bin/env python # -*- encoding: utf-8 -*- import math import torch from torch import Tensor from torch import nn as nn from colossalai import kernel from colossalai import nn as col_nn from colossalai.core import global_context as gpc from colossalai.kernel.cuda_native.scaled_softmax import AttnMaskType from c...
import os from functools import partial from time import time import psutil import torch import torch.nn as nn from commons.model_zoo import model_builder from commons.utils import get_data, get_profile_context, get_tflops, get_time_stamp from packaging import version from torch.nn.parallel import DistributedDataParal...
from torch import nn from transformers import GPT2Config, GPT2LMHeadModel ## Define the Model and Loss Based on Huggingface transformers GPT2LMHeadModel class GPTLMModel(nn.Module): def __init__(self, hidden_size=768, num_layers=12, num_attention_heads=12, ...
import time from contextlib import nullcontext import torch from torch.profiler import ProfilerActivity, profile, schedule, tensorboard_trace_handler class DummyProfiler: def __init__(self): self.step_number = 0 def step(self): self.step_number += 1 # Randomly Generated Data def get_data(...
from torch import nn from transformers import GPT2Config, GPT2LMHeadModel ## Define the Model and Loss Based on Huggingface transformers GPT2LMHeadModel class GPTLMModel(nn.Module): def __init__(self, hidden_size=768, num_layers=12, num_attention_heads=12, ...
import argparse import time from functools import partial import torch from model_zoo import model_builder from torch import nn from tqdm import tqdm from colossalai.fx import ColoTracer from colossalai.fx.passes.adding_split_node_pass import avgnode_split_pass, split_with_split_nodes_pass from colossalai.logging imp...
from typing import Optional, Tuple, Union import torch import torch.nn as nn from transformers.activations import ACT2FN from transformers.models.gpt2.modeling_gpt2 import BaseModelOutputWithPastAndCrossAttentions, GPT2PreTrainedModel from transformers.pytorch_utils import Conv1D class GPT2MLP(nn.Module): def _...
from functools import partial from time import time from typing import Dict, Optional, Tuple, Union import psutil import torch import torch.multiprocessing as mp import torch.nn as nn import transformers from gpt_modules import GPT2LMHeadModel, GPTLMLoss from torch.fx import GraphModule from colossalai.auto_parallel....
import torch # Randomly Generated Data def get_data(batch_size, seq_len, vocab_size): input_ids = torch.randint(0, vocab_size, (batch_size, seq_len), device=torch.cuda.current_device()) attention_mask = torch.ones_like(input_ids) return input_ids, attention_mask def get_tflops(model_numel, batch_size, s...
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
from colossalai.zero.shard_utils import TensorShardStrategy from colossalai.nn.optimizer import FusedAdam # fp16 = dict( # mode=AMP_TYPE.TORCH, # ) # seed = 2 zero = dict(model_config=dict(shard_strategy=TensorShardStrategy(), reduce_scatter_bucket_size_mb=25, ...
from colossalai.zero.shard_utils import TensorShardStrategy from colossalai.nn.optimizer import FusedAdam clip_grad_norm = 1.0
import multiprocessing import os import re from tqdm import tqdm from typing import List import json import time import argparse import functools def split_sentence(document: str, flag: str = "all", limit: int = 510) -> List[str]: """ Args: document: flag: Type:str, "all" 中英文标点分句,"zh" 中文标点分句,"...
import time import os import psutil import h5py import socket import argparse import numpy as np import multiprocessing from tqdm import tqdm from random import shuffle from transformers import AutoTokenizer from get_mask import PreTrainingDataset def get_raw_instance(document, max_sequence_length=512): """ ...
import torch import os from enum import IntEnum from random import choice import random import collections import time import logging import jieba jieba.setLogLevel(logging.CRITICAL) import re import numpy as np import mask PAD = 0 MaskedLMInstance = collections.namedtuple("MaskedLMInstance", ...
class BertDatasetProviderInterface: def get_shard(self, index, shuffle=True): raise NotImplementedError def release_shard(self, index): raise NotImplementedError def prefetch_shard(self, index): raise NotImplementedError def get_batch(self, batch_iter): raise NotImplem...
import os import random import h5py import logging import json import time from concurrent.futures import ProcessPoolExecutor import numpy as np import torch import torch.distributed as dist from torch.utils.data import DataLoader, Dataset from torch.utils.data.sampler import RandomSampler from torch.utils.data.distr...
import colossalai from numpy import require __all__ = ['parse_args'] def parse_args(): parser = colossalai.get_default_parser() parser.add_argument( '--lr', type=float, required=True, help='initial learning rate') parser.add_argument( '--epoch', typ...
import os import math import torch from tqdm import tqdm from utils.global_vars import get_timers, get_tensorboard_writer from nvidia_bert_dataset_provider import NvidiaBertDatasetProvider def evaluate(engine, args, logger, global_step): evaluate_dataset_provider = NvidiaBertDatasetProvider(args, evaluate=True) ...
import transformers import logging from colossalai.nn.lr_scheduler import LinearWarmupLR from transformers import get_linear_schedule_with_warmup from transformers import BertForPreTraining, RobertaForMaskedLM, RobertaConfig from transformers import GPT2Config, GPT2LMHeadModel from transformers import AutoTokenizer, Au...
import colossalai import math import torch from colossalai.context import ParallelMode from colossalai.core import global_context as gpc import colossalai.nn as col_nn from arguments import parse_args from pretrain_utils import get_model, get_optimizer, get_lr_scheduler, save_ckpt from utils.exp_util import get_tflops,...
import torch __all__ = ['LossForPretraining'] class LossForPretraining(torch.nn.Module): def __init__(self, vocab_size): super(LossForPretraining, self).__init__() self.loss_fn = torch.nn.CrossEntropyLoss(ignore_index=-1) self.vocab_size = vocab_size def forward(self, prediction_sco...
import functools import os, shutil import torch import psutil from colossalai.core import global_context as gpc def logging(s, log_path, print_=True, log_=True): if print_: print(s) if log_: with open(log_path, 'a+') as f_log: f_log.write(s + '\n') def get_logger(log_path, **kwargs...
import time import wandb import os from torch.utils.tensorboard import SummaryWriter class WandbLog: @classmethod def init_wandb(cls, project, notes=None, name=time.strftime("%Y-%m-%d %H:%M:%S", time.localtime()), config=None): wandb.init(project=project, notes=notes, name=name, config=config) @c...
import os import logging import torch.distributed as dist logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO) logger = logging.getLogger(__name__) class Logger(): def __init__(self, log_path, cuda=False, debug=False): ...
import time import torch from .WandbLog import TensorboardLog _GLOBAL_TIMERS = None _GLOBAL_TENSORBOARD_WRITER = None def set_global_variables(launch_time, tensorboard_path): _set_timers() _set_tensorboard_writer(launch_time, tensorboard_path) def _set_timers(): """Initialize timers.""" global _GLOB...
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
# coding=utf-8 # Copyright 2020 Microsoft and the Hugging Face Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
import gzip import random from functools import partial from time import time import numpy as np import torch import torch.nn as nn import torch.optim as optim import tqdm from packaging import version from palm_pytorch import PaLM from palm_pytorch.autoregressive_wrapper import AutoregressiveWrapper from torch.utils....
import torch import torch.nn.functional as F from einops import rearrange from torch import einsum, matmul, nn # normalization # they use layernorm without bias, something that pytorch does not offer class LayerNorm(nn.Module): def __init__(self, dim, eps=1e-5): super().__init__() self.eps = eps...
import torch import torch.nn.functional as F from einops import rearrange from torch import nn # helper function def exists(val): return val is not None def eval_decorator(fn): def inner(model, *args, **kwargs): was_training = model.training model.eval() out = fn(model, *args, **kw...
from palm_pytorch.palm_pytorch import PaLM
#!/usr/bin/env python # coding: utf-8 import argparse import os import re import requests COMMIT_API = 'https://api.github.com/repos/hpcaitech/ColossalAI/commits' TAGS_API = 'https://api.github.com/repos/hpcaitech/ColossalAI/tags' def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('--...
import argparse import os def compare_dirs(dir1, dir2): # First, we need to check if the two directories exist if not os.path.exists(dir1) or not os.path.exists(dir2): return False # Now, we compare the list of items in each directory items1 = os.listdir(dir1) items2 = os.listdir(dir2) ...
import os from dataclasses import dataclass from datetime import datetime, timedelta from typing import Any, Dict, List import matplotlib.pyplot as plt import pytz import requests import seaborn from requests_toolbelt import MultipartEncoder @dataclass class Contributor: """ Dataclass for a github contributo...
import argparse import requests def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('-m', '--message', type=str) parser.add_argument('-u', '--url', type=str) return parser.parse_args() def send_message_to_lark(message, webhook_url): data = {"msg_type": "text", "content": {"...
import argparse import os def check_inputs(input_list): for path in input_list: real_path = os.path.join('examples', path) if not os.path.exists(real_path): return False return True def main(): parser = argparse.ArgumentParser() parser.add_argument('-f', '--fileNameList',...
import argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-f', '--fileNameList', type=str, help="The list of changed files") args = parser.parse_args() name_list = args.fileNameList.split(":") folder_need_check = set() for loc in name_list: # Find only the sub...
import os def show_files(path, all_files): # Traverse all the folder/file in current directory file_list = os.listdir(path) # Determine the element is folder or file. If file, pass it into list, if folder, recurse. for file_name in file_list: # Get the abs directory using os.path.join() and st...
from setuptools import find_packages, setup def fetch_requirements(path): with open(path, 'r') as fd: return [r.strip() for r in fd.readlines()] def fetch_readme(): with open('README.md', encoding='utf-8') as f: return f.read() def fetch_version(): with open('version.txt', 'r') as f: ...
import os from copy import deepcopy from functools import partial import pytest import torch import torch.distributed as dist import torch.multiprocessing as mp from chatgpt.experience_maker import NaiveExperienceMaker from chatgpt.nn import GPTActor, GPTCritic, RewardModel from chatgpt.replay_buffer import NaiveRepla...
import argparse from copy import deepcopy import pandas as pd from chatgpt.nn import BLOOMActor, BLOOMCritic, GPTActor, GPTCritic, OPTActor, OPTCritic, RewardModel from chatgpt.trainer import PPOTrainer from chatgpt.trainer.strategies import ColossalAIStrategy, DDPStrategy, NaiveStrategy from torch.optim import Adam f...
import argparse from copy import deepcopy import torch from chatgpt.nn import BLOOMActor, BLOOMCritic, GPTActor, GPTCritic, OPTActor, OPTCritic, RewardModel from chatgpt.trainer import PPOTrainer from chatgpt.trainer.strategies import ColossalAIStrategy, DDPStrategy, NaiveStrategy from torch.optim import Adam from tra...
import argparse import loralib as lora import torch from chatgpt.dataset import RewardDataset from chatgpt.nn import BLOOMRM from chatgpt.trainer import RewardModelTrainer from chatgpt.trainer.strategies import ColossalAIStrategy, DDPStrategy, NaiveStrategy from datasets import load_dataset from torch.optim import Ada...
import argparse from copy import deepcopy import torch import torch.distributed as dist import torch.nn as nn from chatgpt.nn import OPTActor, OPTCritic, RewardModel from chatgpt.trainer import PPOTrainer from chatgpt.trainer.callbacks import PerformanceEvaluator from chatgpt.trainer.strategies import ColossalAIStrate...
import argparse from copy import deepcopy import torch import torch.distributed as dist import torch.nn as nn from chatgpt.nn import GPTActor, GPTCritic, RewardModel from chatgpt.trainer import PPOTrainer from chatgpt.trainer.callbacks import PerformanceEvaluator from chatgpt.trainer.strategies import ColossalAIStrate...
from .base import ReplayBuffer from .naive import NaiveReplayBuffer __all__ = ['ReplayBuffer', 'NaiveReplayBuffer']
import random from typing import List import torch from chatgpt.experience_maker.base import Experience from .base import ReplayBuffer from .utils import BufferItem, make_experience_batch, split_experience_batch class NaiveReplayBuffer(ReplayBuffer): """Naive replay buffer class. It stores experience. Arg...
from dataclasses import dataclass from typing import List, Optional import torch import torch.nn.functional as F from chatgpt.experience_maker.base import Experience @dataclass class BufferItem: """BufferItem is an item of experience data. Shapes of each tensor: sequences: (S) action_log_probs: (A) ...