gatmiry's picture
Upload SortGPT checkpoints (8 configs Γ— 5 seeds Γ— 20 ckpts)
ce004db verified
"""
Shared SortGPT model definition and data utilities.
All experiment scripts import from here to avoid duplication.
"""
import math
from contextlib import nullcontext, contextmanager
from dataclasses import dataclass
import torch
import torch.nn as nn
import torch.nn.functional as F
# ── Device setup ──────────────────────────────────────────────────────────────
DEVICE = torch.device("cuda" if torch.cuda.is_available() else "cpu")
if DEVICE.type == "cuda":
torch.backends.cuda.matmul.allow_tf32 = True
try:
torch.set_float32_matmul_precision("high")
except Exception:
pass
try:
BF16_OK = bool(torch.cuda.is_available() and torch.cuda.is_bf16_supported())
except Exception:
BF16_OK = False
AMP_DTYPE = torch.bfloat16 if BF16_OK else torch.float16
# ── Helpers ───────────────────────────────────────────────────────────────────
def make_generator(device, seed):
try:
g = torch.Generator(device=device.type)
except Exception:
g = torch.Generator()
g.manual_seed(int(seed))
return g
def autocast_ctx(device, enabled=True):
if (not enabled) or device.type != "cuda":
return nullcontext()
try:
return torch.amp.autocast("cuda", dtype=AMP_DTYPE)
except Exception:
return torch.cuda.amp.autocast(dtype=AMP_DTYPE)
def make_grad_scaler(enabled):
if not enabled:
class _NoScaler:
def is_enabled(self): return False
def scale(self, x): return x
def step(self, opt): opt.step()
def update(self): pass
def unscale_(self, opt): pass
return _NoScaler()
try:
return torch.amp.GradScaler()
except Exception:
return torch.cuda.amp.GradScaler()
def float_token(value):
"""Encode a float for use in filenames: 0.02 -> '0p02', -0.1 -> 'm0p1'."""
return str(value).replace("-", "m").replace(".", "p")
# ── Data generation ──────────────────────────────────────────────────────────
def _sample_numbers(batch_size, vocab_n, length, device, allow_duplicates, *, generator=None):
if allow_duplicates:
return torch.randint(0, vocab_n, (batch_size, length), device=device,
generator=generator, dtype=torch.long)
scores = torch.rand(batch_size, vocab_n, device=device, generator=generator)
return scores.topk(length, dim=1).indices.to(torch.long)
def get_batch(batch_size, length, device, *, vocab_n, allow_duplicates=False, generator=None):
"""
Generate a batch for the sorting task.
Returns tensor of shape (batch_size, 2*length+1):
[unsorted_tokens | SEP | sorted_tokens]
SEP token = vocab_n (one above the max token value).
"""
x = _sample_numbers(batch_size, vocab_n, length, device, allow_duplicates, generator=generator)
vals = x.sort(dim=1).values
sep = torch.full((batch_size, 1), vocab_n, device=device, dtype=torch.long)
return torch.cat([x, sep, vals], dim=1)
# ── Model ─────────────────────────────────────────────────────────────────────
class MLP(nn.Module):
def __init__(self, n_embd):
super().__init__()
self.fc_1 = nn.Linear(n_embd, 3 * n_embd)
self.gelu = nn.GELU(approximate="tanh")
self.fc_2 = nn.Linear(3 * n_embd, n_embd)
def forward(self, x):
return self.fc_2(self.gelu(self.fc_1(x)))
class CausalSelfAttention(nn.Module):
def __init__(self, n_embd, n_heads, n_layers):
super().__init__()
assert n_embd % n_heads == 0
self.n_embd = int(n_embd)
self.n_heads = int(n_heads)
self.head_dim = int(n_embd // n_heads)
self.c_attn = nn.Linear(n_embd, 3 * n_embd)
self.c_proj = nn.Linear(n_embd, n_embd)
def forward(self, x):
B, T, C = x.size()
qkv = self.c_attn(x)
q, k, v = qkv.split(self.n_embd, dim=2)
q = q.view(B, T, self.n_heads, self.head_dim).transpose(1, 2)
k = k.view(B, T, self.n_heads, self.head_dim).transpose(1, 2)
v = v.view(B, T, self.n_heads, self.head_dim).transpose(1, 2)
y = F.scaled_dot_product_attention(q, k, v, dropout_p=0.0, is_causal=True)
y = y.transpose(1, 2).contiguous().view(B, T, C)
return self.c_proj(y)
class Block(nn.Module):
def __init__(self, n_embd, n_heads, n_layers, use_mlp=True):
super().__init__()
self.attn = CausalSelfAttention(n_embd, n_heads, n_layers)
self.ln_1 = nn.LayerNorm(n_embd)
self.use_mlp = bool(use_mlp)
if self.use_mlp:
self.mlp = MLP(n_embd)
self.ln_2 = nn.LayerNorm(n_embd)
else:
self.mlp = None
self.ln_2 = None
def forward(self, x):
x = x + self.attn(self.ln_1(x))
if self.mlp is not None:
x = x + self.mlp(self.ln_2(x))
return x
@dataclass
class GPTConfig:
block_size: int
vocab_size: int
n_layers: int
n_heads: int
n_embd: int
without_pos: bool
use_mlp: bool
use_final_LN: bool
max_seq_len: int
class GPT(nn.Module):
_init_std = 0.02 # Set before __init__ to control initialization scale
def __init__(self, config):
super().__init__()
self.config = config
self.n_layers = int(config.n_layers)
self.transformer = nn.ModuleDict(dict(
wte=nn.Embedding(config.vocab_size, config.n_embd),
wpe=nn.Embedding(config.max_seq_len, config.n_embd),
h=nn.ModuleList([
Block(config.n_embd, config.n_heads, config.n_layers, use_mlp=config.use_mlp)
for _ in range(config.n_layers)
]),
ln_f=(nn.LayerNorm(config.n_embd) if config.use_final_LN else nn.Identity()),
))
self.lm_head = nn.Linear(config.n_embd, config.vocab_size, bias=False)
self.lm_head.weight = self.transformer.wte.weight
self.apply(self._init_weights)
self.register_buffer("pos_idx", torch.arange(config.max_seq_len), persistent=False)
if config.without_pos:
with torch.no_grad():
self.transformer.wpe.weight.zero_()
self.transformer.wpe.weight.requires_grad_(False)
def _init_weights(self, module):
std = self.__class__._init_std
if isinstance(module, nn.Linear):
nn.init.normal_(module.weight, mean=0, std=std)
if module.bias is not None:
nn.init.zeros_(module.bias)
elif isinstance(module, nn.Embedding):
nn.init.normal_(module.weight, mean=0, std=std)
def forward(self, idx, *, block_size, return_full_logits=False):
B, T = idx.size()
expected_T = 2 * int(block_size) + 1
assert T == expected_T, f"Expected T={expected_T}, got T={T}"
assert T <= self.config.max_seq_len, f"T={T} exceeds max_seq_len={self.config.max_seq_len}"
pos = self.transformer.wpe(self.pos_idx[:T])
x = self.transformer.wte(idx) if self.config.without_pos else (self.transformer.wte(idx) + pos)
for block in self.transformer.h:
x = block(x)
x = self.transformer.ln_f(x)
logits_half = self.lm_head(x[:, block_size:T - 1, :])
targets = idx[:, block_size + 1:]
loss = F.cross_entropy(logits_half.reshape(-1, logits_half.size(-1)), targets.reshape(-1))
if return_full_logits:
return self.lm_head(x), loss
return logits_half, loss
# ── Model loading ────────────────────────────────────────────────────────────
def load_model_from_checkpoint(ckpt_path, *, extended_max_seq_len=None):
"""
Load a model from a checkpoint file.
Args:
ckpt_path: Path to .pt checkpoint file.
extended_max_seq_len: If set, extend the positional embedding table
to support longer sequences at eval time. Only works when
without_pos=True (pos embeddings are zeroed).
Returns:
model: GPT model on DEVICE in eval mode.
"""
artifact = torch.load(ckpt_path, map_location="cpu")
cfg_dict = artifact["model_config"]
model_cfg = GPTConfig(**cfg_dict)
model = GPT(model_cfg)
model.load_state_dict(artifact["model_state_dict"])
if extended_max_seq_len and extended_max_seq_len > cfg_dict["max_seq_len"]:
model.config = GPTConfig(**dict(cfg_dict, max_seq_len=extended_max_seq_len))
new_wpe = nn.Embedding(extended_max_seq_len, model_cfg.n_embd)
with torch.no_grad():
new_wpe.weight.zero_()
new_wpe.weight.requires_grad_(False)
model.transformer.wpe = new_wpe
model.register_buffer("pos_idx", torch.arange(extended_max_seq_len), persistent=False)
return model.to(DEVICE).eval()
# ── LR schedule ──────────────────────────────────────────────────────────────
def get_lr(itr, max_iters, learning_rate, warmup_iters, min_lr):
"""Cosine decay with linear warmup."""
if itr < warmup_iters:
return learning_rate * (itr + 1) / (warmup_iters + 1)
if itr >= max_iters:
return min_lr
ratio = (itr - warmup_iters) / max(max_iters - warmup_iters, 1)
ratio = 0.5 * (1.0 + math.cos(math.pi * ratio))
return min_lr + ratio * (learning_rate - min_lr)
def create_optimizer(model, *, weight_decay, lr):
params = [p for p in model.parameters() if p.requires_grad]
if DEVICE.type == "cuda":
try:
return torch.optim.AdamW(params, lr=lr, betas=(0.9, 0.95), eps=1e-8,
weight_decay=float(weight_decay), fused=True)
except Exception:
pass
return torch.optim.AdamW(params, lr=lr, betas=(0.9, 0.95), eps=1e-8,
weight_decay=float(weight_decay))