| | """ |
| | SID-GPT v2 training script. |
| | |
| | nanoGPT-style training loop with frame-aligned batch sampling, |
| | cosine LR schedule, gradient accumulation, and AMP support. |
| | Supports single-GPU and multi-GPU (DDP via torchrun). |
| | """ |
| |
|
| | import argparse |
| | import math |
| | import os |
| | import struct |
| | import time |
| | from contextlib import nullcontext |
| |
|
| | import numpy as np |
| | import torch |
| | import torch.distributed as dist |
| | from torch.nn.parallel import DistributedDataParallel as DDP |
| |
|
| | from model import ModelConfig, Transformer |
| |
|
| | TOKEN_SEP = 256 |
| | TOKEN_FRAME = 257 |
| | TOKENS_PER_FRAME = 26 |
| | BYTES_PER_FRAME = 25 |
| |
|
| |
|
| | def setup_ddp(): |
| | """ |
| | Auto-detect DDP: torchrun sets RANK/LOCAL_RANK env vars. |
| | Returns (rank, local_rank, world_size, is_ddp). |
| | Without torchrun, returns (0, 0, 1, False). |
| | """ |
| | if "RANK" not in os.environ: |
| | return 0, 0, 1, False |
| | rank = int(os.environ["RANK"]) |
| | local_rank = int(os.environ["LOCAL_RANK"]) |
| | world_size = int(os.environ["WORLD_SIZE"]) |
| | torch.cuda.set_device(local_rank) |
| | dist.init_process_group(backend="nccl") |
| | return rank, local_rank, world_size, True |
| |
|
| |
|
| | def get_device(requested: str) -> str: |
| | if requested != "auto": |
| | return requested |
| | if torch.cuda.is_available(): |
| | return "cuda" |
| | if hasattr(torch.backends, "mps") and torch.backends.mps.is_available(): |
| | return "mps" |
| | return "cpu" |
| |
|
| |
|
| | def get_dtype(requested: str, device: str) -> torch.dtype: |
| | if requested == "bfloat16": |
| | if device == "cuda" and torch.cuda.is_bf16_supported(): |
| | return torch.bfloat16 |
| | print("[WARN] bfloat16 not supported, falling back to float16") |
| | return torch.float16 |
| | if requested == "float16": |
| | return torch.float16 |
| | return torch.float32 |
| |
|
| |
|
| | def load_data(path: str, device: str) -> torch.Tensor: |
| | raw = np.fromfile(path, dtype=np.uint16) |
| | print(f"[DATA] Loaded {len(raw)} tokens from {path}") |
| | return torch.from_numpy(raw.astype(np.int64)).to(device) |
| |
|
| |
|
| | def generate_synth_data(device: str) -> torch.Tensor: |
| | """ |
| | Generate synthetic training data: ~20 short songs with |
| | deterministic patterns (ascending frequencies, simple ADSR) |
| | for end-to-end pipeline testing without HVSC data. |
| | """ |
| | tokens = [] |
| | rng = np.random.RandomState(42) |
| |
|
| | for song_idx in range(20): |
| | |
| | tokens.extend([TOKEN_SEP] * TOKENS_PER_FRAME) |
| |
|
| | num_frames = 80 + song_idx * 5 |
| | base_freq = 1000 + song_idx * 200 |
| |
|
| | for f in range(num_frames): |
| | tokens.append(TOKEN_FRAME) |
| | regs = [0] * BYTES_PER_FRAME |
| |
|
| | |
| | freq = (base_freq + f * 50) & 0xFFFF |
| | regs[0] = freq & 0xFF |
| | regs[1] = (freq >> 8) & 0xFF |
| | |
| | regs[2] = 0x00 |
| | regs[3] = 0x08 |
| | |
| | regs[4] = 0x11 if f < num_frames - 5 else 0x10 |
| | |
| | regs[5] = 0x09 |
| | regs[6] = 0x00 |
| |
|
| | |
| | freq2 = (base_freq + f * 37 + 500) & 0xFFFF |
| | regs[7] = freq2 & 0xFF |
| | regs[8] = (freq2 >> 8) & 0xFF |
| | regs[9] = 0x00 |
| | regs[10] = 0x08 |
| | regs[11] = 0x21 if f % 16 < 12 else 0x20 |
| | regs[12] = 0x0A |
| | regs[13] = 0x00 |
| |
|
| | |
| | freq3 = (base_freq // 2 + f * 10) & 0xFFFF |
| | regs[14] = freq3 & 0xFF |
| | regs[15] = (freq3 >> 8) & 0xFF |
| | regs[16] = 0x00 |
| | regs[17] = 0x04 |
| | regs[18] = 0x41 if f % 32 < 24 else 0x40 |
| | regs[19] = 0x0C |
| | regs[20] = 0x00 |
| |
|
| | |
| | regs[21] = 0x00 |
| | regs[22] = rng.randint(0, 8) |
| | regs[23] = 0x00 |
| | regs[24] = 0x0F |
| |
|
| | tokens.extend(regs) |
| |
|
| | data = np.array(tokens, dtype=np.uint16) |
| | print(f"[SYNTH] Generated {len(data)} tokens ({20} songs)") |
| | return torch.from_numpy(data.astype(np.int64)).to(device) |
| |
|
| |
|
| | def split_data(data, block_size): |
| | """Split at frame-aligned boundary (multiple of 26).""" |
| | n = len(data) |
| | split_tok = int(n * 0.95) |
| | |
| | split_tok = (split_tok // TOKENS_PER_FRAME) * TOKENS_PER_FRAME |
| | return data[:split_tok], data[split_tok:] |
| |
|
| |
|
| | def get_batch(data, block_size, batch_size, device): |
| | """ |
| | Frame-aligned batch sampling. Offsets are multiples of 26 |
| | so sequences always start on frame boundaries. |
| | """ |
| | max_start = (len(data) - block_size - 1) // TOKENS_PER_FRAME |
| | if max_start < 1: |
| | max_start = 1 |
| | offsets = torch.randint(max_start, (batch_size,)) * TOKENS_PER_FRAME |
| | x = torch.stack([data[o : o + block_size] for o in offsets]) |
| | y = torch.stack( |
| | [data[o + 1 : o + 1 + block_size] for o in offsets] |
| | ) |
| | return x.to(device), y.to(device) |
| |
|
| |
|
| | @torch.no_grad() |
| | def estimate_loss( |
| | model, train_data, val_data, config, args, device, |
| | ): |
| | model.eval() |
| | out = {} |
| | for name, data in [("train", train_data), ("val", val_data)]: |
| | losses = [] |
| | for _ in range(args.eval_iters): |
| | x, y = get_batch( |
| | data, config.block_size, |
| | args.batch_size, device, |
| | ) |
| | with torch.amp.autocast( |
| | device_type=device.split(":")[0], |
| | dtype=args.amp_dtype, |
| | ): |
| | _, loss = model(x, y) |
| | losses.append(loss.item()) |
| | out[name] = sum(losses) / len(losses) |
| | model.train() |
| | return out |
| |
|
| |
|
| | def get_lr(step, args): |
| | """ |
| | Cosine LR schedule with linear warmup. |
| | Decays from lr to min_lr over max_steps. |
| | """ |
| | if step < args.warmup: |
| | return args.lr * (step + 1) / args.warmup |
| | if step >= args.max_steps: |
| | return args.min_lr |
| | progress = (step - args.warmup) / (args.max_steps - args.warmup) |
| | coeff = 0.5 * (1.0 + math.cos(math.pi * progress)) |
| | return args.min_lr + coeff * (args.lr - args.min_lr) |
| |
|
| |
|
| | def configure_optimizer(model, args, device): |
| | |
| | decay_params = [] |
| | no_decay_params = [] |
| | for name, p in model.named_parameters(): |
| | if not p.requires_grad: |
| | continue |
| | if p.dim() >= 2: |
| | decay_params.append(p) |
| | else: |
| | no_decay_params.append(p) |
| |
|
| | groups = [ |
| | {"params": decay_params, "weight_decay": args.weight_decay}, |
| | {"params": no_decay_params, "weight_decay": 0.0}, |
| | ] |
| |
|
| | use_fused = device.startswith("cuda") |
| | optimizer = torch.optim.AdamW( |
| | groups, |
| | lr=args.lr, |
| | betas=(args.beta1, args.beta2), |
| | fused=use_fused, |
| | ) |
| | return optimizer |
| |
|
| |
|
| | def save_checkpoint(model, optimizer, config, step, path): |
| | torch.save( |
| | { |
| | "model": model.state_dict(), |
| | "optimizer": optimizer.state_dict(), |
| | "config": config, |
| | "step": step, |
| | }, |
| | path, |
| | ) |
| | print(f"[CKPT] Saved {path}") |
| |
|
| |
|
| | def main(): |
| | parser = argparse.ArgumentParser( |
| | description="SID-GPT v2 training" |
| | ) |
| | parser.add_argument("--data", type=str, default=None) |
| | parser.add_argument( |
| | "--config", type=str, default="small", |
| | choices=["small", "large"], |
| | ) |
| | parser.add_argument("--batch-size", type=int, default=8) |
| | parser.add_argument("--grad-accum", type=int, default=4) |
| | parser.add_argument("--max-steps", type=int, default=5000) |
| | parser.add_argument("--lr", type=float, default=3e-4) |
| | parser.add_argument("--min-lr", type=float, default=3e-5) |
| | parser.add_argument("--warmup", type=int, default=200) |
| | parser.add_argument("--weight-decay", type=float, default=0.1) |
| | parser.add_argument("--beta1", type=float, default=0.9) |
| | parser.add_argument("--beta2", type=float, default=0.95) |
| | parser.add_argument("--eval-interval", type=int, default=250) |
| | parser.add_argument("--eval-iters", type=int, default=50) |
| | parser.add_argument("--log-interval", type=int, default=10) |
| | parser.add_argument( |
| | "--out-dir", type=str, default="training/checkpoints" |
| | ) |
| | parser.add_argument("--device", type=str, default="auto") |
| | parser.add_argument( |
| | "--dtype", type=str, default="bfloat16", |
| | choices=["bfloat16", "float16", "float32"], |
| | ) |
| | parser.add_argument("--compile", action="store_true") |
| | parser.add_argument("--seed", type=int, default=1337) |
| | parser.add_argument("--synth", action="store_true") |
| | parser.add_argument("--resume", type=str, default=None) |
| | args = parser.parse_args() |
| |
|
| | if not args.synth and args.data is None and args.resume is None: |
| | parser.error("--data or --synth or --resume required") |
| |
|
| | |
| | os.environ["TORCH_ROCM_AOTRITON_ENABLE_EXPERIMENTAL"] = "1" |
| |
|
| | |
| | rank, local_rank, world_size, is_ddp = setup_ddp() |
| | is_master = rank == 0 |
| |
|
| | if is_ddp: |
| | device = f"cuda:{local_rank}" |
| | device_type = "cuda" |
| | else: |
| | device = get_device(args.device) |
| | device_type = device.split(":")[0] |
| |
|
| | torch.manual_seed(args.seed + rank) |
| | args.amp_dtype = get_dtype(args.dtype, device) |
| |
|
| | if is_master: |
| | if is_ddp: |
| | print( |
| | f"[INIT] DDP: {world_size} GPUs, " |
| | f"dtype: {args.amp_dtype}" |
| | ) |
| | else: |
| | print( |
| | f"[INIT] Device: {device}, " |
| | f"dtype: {args.amp_dtype}" |
| | ) |
| |
|
| | |
| | if args.config == "large": |
| | config = ModelConfig.large() |
| | else: |
| | config = ModelConfig.small() |
| |
|
| | start_step = 0 |
| |
|
| | if args.resume: |
| | if is_master: |
| | print(f"[RESUME] Loading checkpoint {args.resume}") |
| | ckpt = torch.load( |
| | args.resume, map_location=device, |
| | weights_only=False, |
| | ) |
| | config = ckpt["config"] |
| | model = Transformer(config).to(device) |
| | model.load_state_dict(ckpt["model"]) |
| | start_step = ckpt["step"] |
| | if is_master: |
| | print(f"[RESUME] Resuming from step {start_step}") |
| | else: |
| | model = Transformer(config).to(device) |
| |
|
| | if is_master: |
| | print( |
| | f"[MODEL] {args.config}: " |
| | f"{model.count_params():,} params, " |
| | f"{config.n_layer}L/{config.n_head}H/" |
| | f"{config.n_embd}D" |
| | ) |
| |
|
| | if args.compile and device_type == "cuda": |
| | if is_master: |
| | print("[COMPILE] torch.compile enabled") |
| | model = torch.compile(model) |
| |
|
| | |
| | if is_ddp: |
| | model = DDP(model, device_ids=[local_rank]) |
| | raw_model = model.module if is_ddp else model |
| |
|
| | |
| | if args.synth: |
| | data = generate_synth_data(device) |
| | else: |
| | data = load_data(args.data, device) |
| |
|
| | train_data, val_data = split_data(data, config.block_size) |
| | if is_master: |
| | print( |
| | f"[DATA] Train: {len(train_data):,} tokens, " |
| | f"Val: {len(val_data):,} tokens" |
| | ) |
| |
|
| | |
| | optimizer = configure_optimizer(raw_model, args, device) |
| |
|
| | if args.resume and "optimizer" in ckpt: |
| | optimizer.load_state_dict(ckpt["optimizer"]) |
| |
|
| | |
| | use_scaler = args.amp_dtype == torch.float16 |
| | scaler = torch.amp.GradScaler(enabled=use_scaler) |
| |
|
| | if is_master: |
| | os.makedirs(args.out_dir, exist_ok=True) |
| |
|
| | |
| | model.train() |
| | t0 = time.time() |
| |
|
| | for step in range(start_step, args.max_steps): |
| | lr = get_lr(step, args) |
| | for pg in optimizer.param_groups: |
| | pg["lr"] = lr |
| |
|
| | |
| | if ( |
| | step % args.eval_interval == 0 |
| | and step > 0 |
| | and is_master |
| | ): |
| | losses = estimate_loss( |
| | model, train_data, val_data, |
| | config, args, device, |
| | ) |
| | print( |
| | f"[EVAL] step {step}: " |
| | f"train={losses['train']:.4f}, " |
| | f"val={losses['val']:.4f}" |
| | ) |
| | save_checkpoint( |
| | raw_model, optimizer, config, step, |
| | os.path.join( |
| | args.out_dir, f"ckpt_{step}.pt" |
| | ), |
| | ) |
| |
|
| | |
| | optimizer.zero_grad(set_to_none=True) |
| | accum_loss = 0.0 |
| |
|
| | for micro in range(args.grad_accum): |
| | x, y = get_batch( |
| | train_data, config.block_size, |
| | args.batch_size, device, |
| | ) |
| | with torch.amp.autocast( |
| | device_type=device_type, dtype=args.amp_dtype |
| | ): |
| | _, loss = model(x, y) |
| | loss = loss / args.grad_accum |
| |
|
| | accum_loss += loss.item() |
| | scaler.scale(loss).backward() |
| |
|
| | scaler.unscale_(optimizer) |
| | torch.nn.utils.clip_grad_norm_( |
| | model.parameters(), 1.0 |
| | ) |
| | scaler.step(optimizer) |
| | scaler.update() |
| |
|
| | |
| | if step % args.log_interval == 0 and is_master: |
| | dt = time.time() - t0 |
| | t0 = time.time() |
| | if dt > 0 and step > start_step: |
| | ms_per_step = ( |
| | dt / args.log_interval * 1000 |
| | ) |
| | tps = ( |
| | args.batch_size * args.grad_accum |
| | * config.block_size |
| | * args.log_interval |
| | * world_size / dt |
| | ) |
| | else: |
| | ms_per_step = 0 |
| | tps = 0 |
| | print( |
| | f"[TRAIN] step {step:5d} | " |
| | f"loss {accum_loss:.4f} | " |
| | f"lr {lr:.2e} | " |
| | f"{ms_per_step:.0f}ms/step | " |
| | f"{dt:.2f}s/{args.log_interval}steps | " |
| | f"{tps/1e6:.2f}M tok/s" |
| | ) |
| |
|
| | |
| | if is_master: |
| | save_checkpoint( |
| | raw_model, optimizer, config, args.max_steps, |
| | os.path.join( |
| | args.out_dir, f"ckpt_{args.max_steps}.pt" |
| | ), |
| | ) |
| |
|
| | losses = estimate_loss( |
| | model, train_data, val_data, |
| | config, args, device, |
| | ) |
| | print( |
| | f"[DONE] Final: train={losses['train']:.4f}, " |
| | f"val={losses['val']:.4f}" |
| | ) |
| |
|
| | if is_ddp: |
| | dist.destroy_process_group() |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|