| import torch |
| from torch.utils.data import Dataset, DataLoader, DistributedSampler |
| from tqdm import tqdm |
| import os |
| import glob |
| import wandb |
| from datasets import load_dataset,concatenate_datasets |
| import torch.distributed as dist |
| from torch.nn.parallel import DistributedDataParallel as DDP |
| import argparse |
|
|
| |
| from src.rwkv7 import RWKV7 |
| from src.dataset import MyDataset |
| from src.transformer import TransformerModel |
|
|
|
|
| class L2Wrap(torch.autograd.Function): |
| @staticmethod |
| def forward(ctx, loss, y): |
| ctx.save_for_backward(y) |
| return loss |
|
|
| @staticmethod |
| def backward(ctx, grad_output): |
| y = ctx.saved_tensors[0] |
| factor = 1e-4 / (y.shape[0] * y.shape[1]) |
| maxx, ids = torch.max(y, -1, keepdim=True) |
| gy = torch.zeros_like(y) |
| gy.scatter_(-1, ids, maxx * factor) |
| return grad_output, gy |
| |
| def setup(rank, world_size): |
| os.environ['MASTER_ADDR'] = 'localhost' |
| os.environ['MASTER_PORT'] = '12355' |
| dist.init_process_group("nccl", rank=rank, world_size=world_size) |
| torch.cuda.set_device(rank) |
|
|
| def cleanup(): |
| dist.destroy_process_group() |
|
|
| def load_latest_checkpoint(model, checkpoint_dir): |
| checkpoint_files = [f for f in os.listdir(checkpoint_dir) if f.endswith('.pt')] |
| if not checkpoint_files: |
| print("No checkpoint files found in the directory.") |
| return 0 |
| latest_checkpoint = max(checkpoint_files, key=lambda x: os.path.getctime(os.path.join(checkpoint_dir, x))) |
| checkpoint_path = os.path.join(checkpoint_dir, latest_checkpoint) |
| model.load_state_dict(torch.load(checkpoint_path)) |
| print(f"Loaded checkpoint: {checkpoint_path}") |
|
|
| def initialize_model(checkpoint_dir, dim, n_blocks, rank, world_size): |
| model = RWKV7(text_vocab=128, audio_vocab=8192 + 1, dim=dim, n_blocks=n_blocks).to(rank) |
| |
| |
| if rank == 0: |
| load_latest_checkpoint(model, checkpoint_dir) |
| dist.barrier() |
| model = DDP(model, device_ids=[rank], find_unused_parameters=True, broadcast_buffers=False) |
| |
| |
| for param in model.parameters(): |
| dist.broadcast(param.data.clone(), src=0) |
| |
| |
| |
| |
| total_params = sum(p.numel() for p in model.parameters()) |
| trainable_params = sum(p.numel() for p in model.parameters() if p.requires_grad) |
| print(f"Model total parameters: {total_params}") |
| print(f"Model trainable parameters: {trainable_params}") |
| return model |
|
|
| def collate_fn(batch): |
| padding_token = 8192 |
| max_length = max(len(seq) for seq in batch) - 1 |
|
|
| chunck_length = ((max_length // 16) + 1) * 16 |
| |
| input_ids, targets, loss_masks = [], [], [] |
| for seq in batch: |
| input_seq = list(seq[:-1]) |
| target_seq = list(seq[1:]) |
| input_padding = [padding_token] * (max_length - len(input_seq)) |
| target_padding = [padding_token] * (max_length - len(target_seq)) |
| mask_padding = [0] * (max_length - len(input_seq)) |
| input_ids.append(torch.tensor(input_seq + input_padding + [padding_token] * (chunck_length - max_length), dtype=torch.long)) |
| targets.append(torch.tensor(target_seq + target_padding + [padding_token] * (chunck_length - max_length), dtype=torch.long)) |
| loss_masks.append(torch.tensor([1] * len(input_seq) + mask_padding + [0] * (chunck_length - max_length), dtype=torch.long)) |
| |
| |
| |
| |
| |
| return torch.stack(input_ids, dim=0), torch.stack(targets, dim=0), torch.stack(loss_masks, dim=0) |
|
|
| def prepare_dataloader(batch_size, rank, world_size): |
| |
| |
| |
| |
| |
| |
| dataset2 = load_dataset("JerryAGENDD/libritts_tokenized_960", cache_dir="../temp_datasets")['train'] |
| dataset = dataset2.remove_columns(['text_original', 'speaker_id']) |
| |
| |
| |
| |
| dataset = MyDataset(hf_dataset=dataset, train_type='pretrain') |
| sampler = DistributedSampler(dataset, num_replicas=world_size, rank=rank,shuffle=True) |
| dataloader = DataLoader(dataset, batch_size=batch_size, sampler=sampler, collate_fn=collate_fn) |
| return dataloader |
|
|
| def train(rank, world_size, args): |
| setup(rank, world_size) |
| model = initialize_model(args.checkpoint_dir, args.dim, args.n_blocks, rank,world_size) |
| dataloader = prepare_dataloader(args.batch_size, rank, world_size) |
| |
| optimizer = torch.optim.Adam(model.parameters(), lr=args.learning_rate) |
| if rank == 0: |
| wandb.init(project="TTS") |
| |
| logging_parameter = True |
| |
| os.makedirs(args.checkpoint_dir, exist_ok=True) |
| model.train() |
| for epoch in range(args.num_epochs): |
| dataloader.sampler.set_epoch(epoch) |
| if rank == 0: |
| epoch_iterator = tqdm(dataloader, desc=f"Epoch {epoch + 1}/{args.num_epochs}", leave=False) |
| else: |
| epoch_iterator = dataloader |
| |
| for batch in epoch_iterator: |
| input_ids, targets, loss_masks = batch |
| input_ids = input_ids.long().to(rank) |
| targets = targets.long().to(rank) |
| loss_masks = loss_masks.to(rank) |
| outputs = model(None, None, input_ids) |
| criterion = torch.nn.CrossEntropyLoss(reduction='none') |
| loss = criterion(outputs.view(-1, outputs.size(-1)), targets.view(-1)) |
| loss = loss.view(targets.size()) * loss_masks |
| loss = loss.sum() / loss_masks.sum() |
| loss = L2Wrap.apply(loss, outputs) |
| |
| if rank == 0: |
| wandb.log({"loss": loss.item()}) |
| |
| optimizer.zero_grad() |
| loss.backward() |
| |
| if(logging_parameter and rank == 0): |
| for name, param in model.named_parameters(): |
| if param.grad is None: |
| print(f"Parameter {name} did not receive gradient") |
|
|
| logging_parameter = False |
| |
| optimizer.step() |
| |
| if rank == 0: |
| save_checkpoint(model, args.checkpoint_dir, epoch) |
| |
| if rank == 0: |
| wandb.finish() |
| |
| cleanup() |
|
|
| def save_checkpoint(model, output_dir, epoch): |
| pt_files = glob.glob(os.path.join(output_dir, "*.pt")) |
| for pt_file in pt_files: |
| os.remove(pt_file) |
| checkpoint_path = os.path.join(output_dir, f"checkpoint_epoch_{epoch + 1}.pt") |
| torch.save(model.module.state_dict(), checkpoint_path) |
|
|
| def main(): |
| parser = argparse.ArgumentParser(description="Train RWKV7 model") |
| parser.add_argument("--dim", type=int, default=128, help="Dimension of the model") |
| parser.add_argument("--n_blocks", type=int, default=5, help="Number of blocks in the model") |
| parser.add_argument("--num_epochs", type=int, default=4000, help="Number of training epochs") |
| parser.add_argument("--learning_rate", type=float, default=1e-4, help="Learning rate for optimizer") |
| parser.add_argument("--batch_size", type=int, default=128, help="Batch size for training") |
| parser.add_argument("--checkpoint_dir", type=str, default="./checkpoints", help="Directory to save checkpoints") |
| args = parser.parse_args() |
| world_size = torch.cuda.device_count() |
| torch.multiprocessing.spawn(train, args=(world_size, args), nprocs=world_size, join=True) |
|
|
| if __name__ == "__main__": |
| main() |