| | import functools |
| | import math |
| | from typing import Callable |
| |
|
| | import einops |
| | import jax |
| | import jax.ad_checkpoint as adc |
| | import jax.numpy as jnp |
| | import ninjax as nj |
| | import numpy as np |
| |
|
| | COMPUTE_DTYPE = jnp.bfloat16 |
| | LAYER_CALLBACK = lambda tensor, name: tensor |
| |
|
| | f32 = jnp.float32 |
| |
|
| |
|
| | def cast(xs, force=False): |
| | if force: |
| | should = lambda x: True |
| | else: |
| | should = lambda x: jnp.issubdtype(x.dtype, jnp.floating) |
| | return jax.tree.map(lambda x: COMPUTE_DTYPE(x) if should(x) else x, xs) |
| |
|
| |
|
| | def act(name): |
| | if name == 'none': |
| | return lambda x: x |
| | elif name == 'mish': |
| | return lambda x: x * jnp.tanh(jax.nn.softplus(x)) |
| | elif name == 'relu2': |
| | return lambda x: jnp.square(jax.nn.relu(x)) |
| | elif name == 'swiglu': |
| | def fn(x): |
| | x, y = jnp.split(x, 2, -1) |
| | return jax.nn.silu(x) * y |
| | return fn |
| | else: |
| | return getattr(jax.nn, name) |
| |
|
| |
|
| | def init(name): |
| | if callable(name): |
| | return name |
| | elif name.endswith(('_in', '_out', '_avg')): |
| | dist, fan = name.rsplit('_', 1) |
| | else: |
| | dist, fan = name, 'in' |
| | return Initializer(dist, fan, 1.0) |
| |
|
| |
|
| | def dropout(x, prob, training): |
| | if not prob or not training: |
| | return x |
| | keep = jax.random.bernoulli(nj.seed(), 1.0 - prob, x.shape) |
| | return x * keep / (1.0 - prob) |
| |
|
| |
|
| | def symlog(x): |
| | return jnp.sign(x) * jnp.log1p(jnp.abs(x)) |
| |
|
| |
|
| | def symexp(x): |
| | return jnp.sign(x) * jnp.expm1(jnp.abs(x)) |
| |
|
| |
|
| | def where(condition, xs, ys): |
| | assert condition.dtype == bool, condition.dtype |
| | def fn(x, y): |
| | assert x.shape == y.shape, (x.shape, y.shape) |
| | expanded = jnp.expand_dims(condition, list(range(condition.ndim, x.ndim))) |
| | return jnp.where(expanded, x, y) |
| | return jax.tree.map(fn, xs, ys) |
| |
|
| |
|
| | def mask(xs, mask): |
| | return where(mask, xs, jax.tree.map(jnp.zeros_like, xs)) |
| |
|
| |
|
| | def available(*trees, bdims=None): |
| | def fn(*xs): |
| | masks = [] |
| | for x in xs: |
| | if jnp.issubdtype(x.dtype, jnp.floating): |
| | mask = (x != -jnp.inf) |
| | elif jnp.issubdtype(x.dtype, jnp.signedinteger): |
| | mask = (x != -1) |
| | elif ( |
| | jnp.issubdtype(x.dtype, jnp.unsignedinteger) or |
| | jnp.issubdtype(x.dtype, bool)): |
| | shape = x.shape if bdims is None else x.shape[:bdims] |
| | mask = jnp.full(shape, True, bool) |
| | else: |
| | raise NotImplementedError(x.dtype) |
| | if bdims is not None: |
| | mask = mask.all(tuple(range(bdims, mask.ndim))) |
| | masks.append(mask) |
| | return jnp.stack(masks, 0).all(0) |
| | return jax.tree.map(fn, *trees) |
| |
|
| |
|
| | @functools.partial(jax.custom_vjp, nondiff_argnums=[1, 2]) |
| | def ensure_dtypes(x, fwd=None, bwd=None): |
| | fwd = fwd or COMPUTE_DTYPE |
| | bwd = bwd or COMPUTE_DTYPE |
| | assert x.dtype == fwd, (x.dtype, fwd) |
| | return x |
| | def ensure_dtypes_fwd(x, fwd=None, bwd=None): |
| | fwd = fwd or COMPUTE_DTYPE |
| | bwd = bwd or COMPUTE_DTYPE |
| | return ensure_dtypes(x, fwd, bwd), () |
| | def ensure_dtypes_bwd(fwd, bwd, cache, dx): |
| | fwd = fwd or COMPUTE_DTYPE |
| | bwd = bwd or COMPUTE_DTYPE |
| | assert dx.dtype == bwd, (dx.dtype, bwd) |
| | return (dx,) |
| | ensure_dtypes.defvjp(ensure_dtypes_fwd, ensure_dtypes_bwd) |
| |
|
| |
|
| | def rms(xs): |
| | xs = jax.tree.leaves(xs) |
| | count = sum(x.size for x in xs) |
| | sumsq = jnp.stack([f32(jnp.square(x).sum()) for x in xs]).sum() |
| | return jnp.sqrt(sumsq / f32(count)) |
| |
|
| |
|
| | def rope(x, ts=None, inverse=False, maxlen=4096): |
| | B, T, _, D = x.shape |
| | if ts is None: |
| | ts = jnp.ones(B, jnp.int32)[:, None] * jnp.arange(T)[None, :] |
| | assert ts.shape == (B, T), (ts.shape, (B, T)) |
| | if inverse: |
| | ts = -ts |
| | freq_exponents = (2.0 / D) * jnp.arange(D // 2) |
| | timescale = maxlen ** freq_exponents |
| | radians = ts[:, :, None] / timescale[None, None, :] |
| | radians = radians[..., None, :].astype(x.dtype) |
| | sin, cos = jnp.sin(radians), jnp.cos(radians) |
| | x1, x2 = jnp.split(x, 2, axis=-1) |
| | res = jnp.concatenate([x1 * cos - x2 * sin, x2 * cos + x1 * sin], axis=-1) |
| | return res |
| |
|
| |
|
| | class Initializer: |
| |
|
| | def __init__(self, dist='trunc_normal', fan='in', scale=1.0): |
| | self.dist = dist |
| | self.fan = fan |
| | self.scale = scale |
| |
|
| | def __call__(self, shape, dtype=jnp.float32, fshape=None): |
| | shape = (shape,) if isinstance(shape, int) else tuple(shape) |
| | assert all(isinstance(x, int) for x in shape), ( |
| | shape, [type(x) for x in shape]) |
| | assert all(x > 0 for x in shape), shape |
| | fanin, fanout = self.compute_fans(shape if fshape is None else fshape) |
| | fan = { |
| | 'avg': (fanin + fanout) / 2, 'in': fanin, 'out': fanout, 'none': 1, |
| | }[self.fan] |
| | if self.dist == 'zeros': |
| | x = jnp.zeros(shape, dtype) |
| | elif self.dist == 'uniform': |
| | limit = np.sqrt(1 / fan) |
| | x = jax.random.uniform(nj.seed(), shape, dtype, -limit, limit) |
| | elif self.dist == 'normal': |
| | x = jax.random.normal(nj.seed(), shape) |
| | x *= np.sqrt(1 / fan) |
| | elif self.dist == 'trunc_normal': |
| | x = jax.random.truncated_normal(nj.seed(), -2, 2, shape) |
| | x *= 1.1368 * np.sqrt(1 / fan) |
| | elif self.dist == 'normed': |
| | x = jax.random.uniform(nj.seed(), shape, dtype, -1, 1) |
| | x *= (1 / jnp.linalg.norm(x.reshape((-1, shape[-1])), 2, 0)) |
| | else: |
| | raise NotImplementedError(self.dist) |
| | x *= self.scale |
| | x = x.astype(dtype) |
| | return x |
| |
|
| | def __repr__(self): |
| | return f'Initializer({self.dist}, {self.fan}, {self.scale})' |
| |
|
| | def __eq__(self, other): |
| | attributes = ('dist', 'fan', 'scale') |
| | return all(getattr(self, k) == getattr(other, k) for k in attributes) |
| |
|
| | @staticmethod |
| | def compute_fans(shape): |
| | if len(shape) == 0: |
| | return (1, 1) |
| | elif len(shape) == 1: |
| | return (1, shape[0]) |
| | elif len(shape) == 2: |
| | return shape |
| | else: |
| | space = math.prod(shape[:-2]) |
| | return (shape[-2] * space, shape[-1] * space) |
| |
|
| |
|
| | class Embed(nj.Module): |
| |
|
| | einit: str | Callable = Initializer('trunc_normal', 'out') |
| | combine: bool = False |
| |
|
| | def __init__(self, classes, units, shape=()): |
| | self.classes = classes |
| | self.units = units |
| | self.shape = shape |
| |
|
| | def __call__(self, x): |
| | batch_shape = x.shape[:x.ndim - len(self.shape)] |
| | event_shape = x.shape[x.ndim - len(self.shape):] |
| | assert event_shape == self.shape, (self.shape, x.shape) |
| | N = math.prod(self.shape) |
| | K = self.classes |
| | D = self.units |
| | shape = (*self.shape, self.classes, self.units) |
| | table = self.value('table', init(self.einit), shape) |
| | table = table.reshape(N, K, D) |
| | table = table.astype(COMPUTE_DTYPE) |
| | index = x.reshape(-1, N) |
| | embed = table[jnp.arange(N), index] |
| | if self.combine: |
| | embed = embed.sum(-2).reshape(*batch_shape, self.units) |
| | else: |
| | embed = embed.reshape(*batch_shape, *self.shape, self.units) |
| | return embed |
| |
|
| |
|
| | class Linear(nj.Module): |
| |
|
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | outscale: float = 1.0 |
| |
|
| | def __init__(self, units): |
| | self.units = (units,) if isinstance(units, int) else tuple(units) |
| |
|
| | def __call__(self, x): |
| | ensure_dtypes(x) |
| | size = math.prod(self.units) |
| | shape = (x.shape[-1], size) |
| | x = x @ self.value('kernel', self._scaled_winit, shape).astype(x.dtype) |
| | if self.bias: |
| | x += self.value('bias', init(self.binit), size).astype(x.dtype) |
| | x = x.reshape((*x.shape[:-1], *self.units)) |
| | return x |
| |
|
| | def _scaled_winit(self, *args, **kwargs): |
| | return init(self.winit)(*args, **kwargs) * self.outscale |
| |
|
| |
|
| | class BlockLinear(nj.Module): |
| |
|
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | outscale: float = 1.0 |
| |
|
| | def __init__(self, units, blocks): |
| | assert isinstance(units, int), (units, type(units)) |
| | assert blocks <= units and units % blocks == 0, (blocks, units) |
| | self.units = units |
| | self.blocks = blocks |
| |
|
| | def __call__(self, x): |
| | ensure_dtypes(x) |
| | assert x.shape[-1] % self.blocks == 0, (x.shape, self.blocks) |
| | insize = x.shape[-1] |
| | shape = (self.blocks, insize // self.blocks, self.units // self.blocks) |
| | kernel = self.value('kernel', self._scaled_winit, shape).astype(x.dtype) |
| | x = x.reshape((*x.shape[:-1], self.blocks, insize // self.blocks)) |
| | x = jnp.einsum('...ki,kio->...ko', x, kernel) |
| | x = x.reshape((*x.shape[:-2], self.units)) |
| | if self.bias: |
| | x += self.value('bias', init(self.binit), self.units).astype(x.dtype) |
| | return x |
| |
|
| | def _scaled_winit(self, *args, **kwargs): |
| | return init(self.winit)(*args, **kwargs) * self.outscale |
| |
|
| |
|
| | class Conv2D(nj.Module): |
| |
|
| | transp: bool = False |
| | groups: int = 1 |
| | pad: str = 'same' |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | outscale: float = 1.0 |
| |
|
| | def __init__(self, depth, kernel, stride=1): |
| | self.depth = depth |
| | self.kernel = (kernel,) * 2 if isinstance(kernel, int) else kernel |
| | self.stride = stride |
| |
|
| | def __call__(self, x): |
| | ensure_dtypes(x) |
| | shape = (*self.kernel, x.shape[-1] // self.groups, self.depth) |
| | kernel = self.value('kernel', self._scaled_winit, shape).astype(x.dtype) |
| | if self.transp: |
| | assert self.pad == 'same', self.pad |
| | |
| | |
| | x = x.repeat(self.stride, -2).repeat(self.stride, -3) |
| | maskh = ((jnp.arange(x.shape[-3]) - 1) % self.stride == 0)[:, None] |
| | maskw = ((jnp.arange(x.shape[-2]) - 1) % self.stride == 0)[None, :] |
| | x *= (maskh * maskw)[:, :, None] |
| | stride = (1, 1) |
| | else: |
| | stride = (self.stride, self.stride) |
| | x = jax.lax.conv_general_dilated( |
| | x, kernel, stride, self.pad.upper(), |
| | feature_group_count=self.groups, |
| | dimension_numbers=('NHWC', 'HWIO', 'NHWC')) |
| | if self.bias: |
| | x += self.value('bias', init(self.binit), self.depth).astype(x.dtype) |
| | return x |
| |
|
| | def _scaled_winit(self, *args, **kwargs): |
| | return init(self.winit)(*args, **kwargs) * self.outscale |
| |
|
| |
|
| | class Conv3D(nj.Module): |
| |
|
| | transp: bool = False |
| | groups: int = 1 |
| | pad: str = 'same' |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| |
|
| | def __init__(self, depth, kernel, stride=1): |
| | self.depth = depth |
| | self.kernel = (kernel,) * 3 if isinstance(kernel, int) else kernel |
| | self.stride = (stride,) * 3 if isinstance(stride, int) else stride |
| |
|
| | def __call__(self, x): |
| | ensure_dtypes(x) |
| | if self.transp: |
| | assert self.groups == 1, self.groups |
| | shape = (*self.kernel, x.shape[-1], self.depth) |
| | kernel = self.value('kernel', init(self.winit), shape).astype(x.dtype) |
| | x = jax.lax.conv_transpose( |
| | x, kernel, self.stride, self.pad.upper(), |
| | dimension_numbers=('NTHWC', 'THWIO', 'NTHWC')) |
| | else: |
| | shape = (*self.kernel, x.shape[-1] // self.groups, self.depth) |
| | kernel = self.value('kernel', init(self.winit), shape).astype(x.dtype) |
| | x = jax.lax.conv_general_dilated( |
| | x, kernel, self.stride, self.pad.upper(), |
| | feature_group_count=self.groups, |
| | dimension_numbers=('NTHWC', 'THWIO', 'NTHWC')) |
| | if self.bias: |
| | x += self.value('bias', init(self.binit), self.depth).astype(x.dtype) |
| | return x |
| |
|
| |
|
| | class Norm(nj.Module): |
| |
|
| | axis: tuple = (-1,) |
| | eps: float = 1e-4 |
| | scale: bool = True |
| | shift: bool = True |
| |
|
| | def __init__(self, impl): |
| | if '1em' in impl: |
| | impl, exp = impl.split('1em') |
| | self._fields['eps'] = 10 ** -int(exp) |
| | self.impl = impl |
| |
|
| | def __call__(self, x): |
| | ensure_dtypes(x) |
| | dtype = x.dtype |
| | x = f32(x) |
| | axis = [a % x.ndim for a in self.axis] |
| | shape = [x.shape[i] if i in axis else 1 for i in range(min(axis), x.ndim)] |
| | if self.impl == 'none': |
| | pass |
| | elif self.impl == 'rms': |
| | mean2 = jnp.square(x).mean(axis, keepdims=True) |
| | mean2 = adc.checkpoint_name(mean2, 'small') |
| | scale = self._scale(shape, x.dtype) |
| | x = x * (jax.lax.rsqrt(mean2 + self.eps) * scale) |
| | elif self.impl == 'layer': |
| | mean = x.mean(axis, keepdims=True) |
| | mean2 = jnp.square(x).mean(axis, keepdims=True) |
| | mean2 = adc.checkpoint_name(mean2, 'small') |
| | var = jnp.maximum(0, mean2 - jnp.square(mean)) |
| | var = adc.checkpoint_name(var, 'small') |
| | scale = self._scale(shape, x.dtype) |
| | shift = self._shift(shape, x.dtype) |
| | x = (x - mean) * (jax.lax.rsqrt(var + self.eps) * scale) + shift |
| | else: |
| | raise NotImplementedError(self.impl) |
| | x = x.astype(dtype) |
| | return x |
| |
|
| | def _scale(self, shape, dtype): |
| | if not self.scale: |
| | return jnp.ones(shape, dtype) |
| | return self.value('scale', jnp.ones, shape, f32).astype(dtype) |
| |
|
| | def _shift(self, shape, dtype): |
| | if not self.shift: |
| | return jnp.zeros(shape, dtype) |
| | return self.value('shift', jnp.zeros, shape, f32).astype(dtype) |
| |
|
| |
|
| | class Attention(nj.Module): |
| |
|
| | heads: int = 8 |
| | kv_heads: int = 0 |
| | dropout: float = 0.0 |
| | rope: bool = True |
| | qknorm: str = 'none' |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | outscale: float = 1.0 |
| |
|
| | def __call__(self, x, mask=None, ts=None, training=True): |
| | kw = dict(bias=self.bias, winit=self.winit, binit=self.binit) |
| | B, T, D = x.shape |
| | kv_heads = self.kv_heads or self.heads |
| | assert self.heads % kv_heads == 0 |
| | head_ratio = self.heads // kv_heads |
| | if head_ratio == 1: |
| | qkv = self.sub('qkv', Linear, 3 * D, **kw)(x) |
| | q, k, v = jnp.split(qkv, 3, -1) |
| | else: |
| | q = self.sub('q', Linear, D, **kw)(x) |
| | k = self.sub('k', Linear, D // head_ratio, **kw)(x) |
| | v = self.sub('v', Linear, D // head_ratio, **kw)(x) |
| | q = einops.rearrange(q, 'b t (h d) -> b t h d', h=self.heads) |
| | k = einops.rearrange(k, 'b t (h d) -> b t h d', h=kv_heads) |
| | v = einops.rearrange(v, 'b t (h d) -> b t h d', h=kv_heads) |
| |
|
| | if self.qknorm != 'none': |
| | q = self.sub('normq', Norm, self.qknorm)(q) |
| | k = self.sub('normk', Norm, self.qknorm)(k) |
| |
|
| | if self.rope: |
| | q = rope(q, ts) |
| | k = rope(k, ts) |
| |
|
| | q = einops.rearrange(q, 'b t (h g) d -> b t h g d', h=kv_heads) |
| | logits = einops.einsum(q, k, 'b tq h g d, b tk h d -> b h g tq tk') |
| | logits = logits * (1.0 / np.sqrt(k.shape[-1])) |
| | logits = f32(logits) |
| | if mask is not None: |
| | Tq, Tk = q.shape[1], k.shape[1] |
| | assert mask.shape == (B, Tq, Tk), (mask.shape, (B, Tq, Tk)) |
| | mask = einops.rearrange(mask, 'b tq tk -> b 1 1 tq tk') |
| | logits = jnp.where(mask, logits, -1e30) |
| | weights = jax.nn.softmax(logits) |
| | weights = weights.astype(x.dtype) |
| | weights = dropout(weights, self.dropout, training) |
| | x = einops.einsum(weights, v, 'b h g tq tk, b tk h d -> b tq h g d') |
| | x = einops.rearrange(x, 'b t h g d -> b t (h g d)') |
| | x = self.sub('proj', Linear, D, **kw, outscale=self.outscale)(x) |
| | return x |
| |
|
| |
|
| | class DictConcat: |
| |
|
| | def __init__(self, spaces, fdims, squish=lambda x: x): |
| | assert 1 <= fdims, fdims |
| | self.keys = sorted(spaces.keys()) |
| | self.spaces = spaces |
| | self.fdims = fdims |
| | self.squish = squish |
| |
|
| | def __call__(self, xs): |
| | assert all(k in xs for k in self.spaces), (self.spaces, xs.keys()) |
| | bdims = xs[self.keys[0]].ndim - len(self.spaces[self.keys[0]].shape) |
| | ys = [] |
| | for key in self.keys: |
| | space = self.spaces[key] |
| | x = xs[key] |
| | m = available(x, bdims=bdims) |
| | x = mask(x, m) |
| | assert x.shape[bdims:] == space.shape, (key, bdims, space.shape, x.shape) |
| | if space.dtype == jnp.uint8 and len(space.shape) in (2, 3): |
| | raise NotImplementedError('Images are not supported.') |
| | elif space.discrete: |
| | classes = np.asarray(space.classes).flatten() |
| | assert (classes == classes[0]).all(), classes |
| | classes = classes[0].item() |
| | x = x.astype(jnp.int32) |
| | x = jax.nn.one_hot(x, classes, dtype=COMPUTE_DTYPE) |
| | else: |
| | x = self.squish(x) |
| | x = x.astype(COMPUTE_DTYPE) |
| | x = mask(x, m) |
| | x = x.reshape((*x.shape[:bdims + self.fdims - 1], -1)) |
| | ys.append(x) |
| | return jnp.concatenate(ys, -1) |
| |
|
| |
|
| | class DictEmbed(nj.Module): |
| |
|
| | squish: Callable = lambda x: x |
| | padone: bool = True |
| | bias: bool = True |
| | einit: str | Callable = Initializer('trunc_normal', 'out') |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | impl: str = 'onehot' |
| |
|
| | def __init__(self, spaces, units): |
| | self.keys = sorted(spaces.keys()) |
| | self.spaces = spaces |
| | self.units = units |
| | self.ekw = dict(einit=self.einit) |
| | self.lkw = dict(bias=self.bias, winit=self.winit, binit=self.binit) |
| |
|
| | def __call__(self, xs, bshape): |
| | assert isinstance(bshape, tuple), bshape |
| | assert all(k in xs for k in self.spaces), (self.spaces, xs.keys()) |
| | ys = [] |
| | init = self.value('init', self.einit, (self.units,)) |
| | init = jnp.broadcast_to(init, (*bshape, self.units)) |
| | init = COMPUTE_DTYPE(init) |
| | ys.append(init) |
| | for key in self.keys: |
| | try: |
| | space = self.spaces[key] |
| | x = xs[key] |
| | assert x.dtype == space.dtype, (key, space.dtype, x.dtype, x.shape) |
| | m = available(x, bdims=len(bshape)) |
| | x = mask(x, m) |
| | if space.discrete: |
| | if space.dtype == jnp.uint8 and len(space.shape) in (2, 3): |
| | raise NotImplementedError('Images are not supported.') |
| | classes = int(np.asarray(space.classes).max()) |
| | assert classes <= 256, (key, space, classes) |
| | if self.impl == 'lookup': |
| | x = self.sub( |
| | key, Embed, classes, self.units, space.shape, |
| | combine=True, **self.ekw)(x) |
| | |
| | elif self.impl == 'onehot': |
| | x = jax.nn.one_hot(x, classes, dtype=COMPUTE_DTYPE) |
| | x = x.reshape((*x.shape[:len(bshape)], -1)) |
| | x = self.sub(key, Linear, self.units, **self.lkw)(x) |
| | else: |
| | raise NotImplementedError(self.impl) |
| | else: |
| | x = self.squish(x) |
| | x = x.astype(COMPUTE_DTYPE) |
| | x = x.reshape((*x.shape[:len(bshape)], -1)) |
| | x = self.sub(key, Linear, self.units, **self.lkw)(x) |
| | x = mask(x, m) |
| | ys.append(x) |
| | except Exception: |
| | print(f"Error encoding key '{key}' with space {space}.") |
| | raise |
| | x = sum(ys) |
| | return x |
| |
|
| |
|
| | class MLP(nj.Module): |
| |
|
| | act: str = 'silu' |
| | norm: str = 'rms' |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| |
|
| | def __init__(self, layers=5, units=1024): |
| | self.layers = layers |
| | self.units = units |
| | self.kw = dict(bias=self.bias, winit=self.winit, binit=self.binit) |
| |
|
| | def __call__(self, x): |
| | shape = x.shape[:-1] |
| | x = x.astype(COMPUTE_DTYPE) |
| | x = x.reshape([-1, x.shape[-1]]) |
| | for i in range(self.layers): |
| | x = self.sub(f'linear{i}', Linear, self.units, **self.kw)(x) |
| | x = self.sub(f'norm{i}', Norm, self.norm)(x) |
| | x = act(self.act)(x) |
| | x = x.reshape((*shape, x.shape[-1])) |
| | return x |
| |
|
| |
|
| | class Transformer(nj.Module): |
| |
|
| | units: int = 1024 |
| | layers: int = 12 |
| | heads: int = 8 |
| | ffup: int = 4 |
| | act: str = 'silu' |
| | norm: str = 'rms' |
| | glu: bool = False |
| | rope: bool = True |
| | qknorm: str = 'none' |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | outscale: float = 1.0 |
| |
|
| | def __call__(self, x, mask=None, ts=None, training=True): |
| | kw = {k: getattr(self, k) for k in ('bias', 'winit', 'binit')} |
| | ak = {k: getattr(self, k) for k in ('heads', 'rope', 'qknorm', 'outscale')} |
| | D = x.shape[-1] |
| | assert D == self.units, (D, self.units) |
| | for i in range(self.layers): |
| | with nj.scope(f'layer{i}'): |
| | skip = x |
| | x = self.sub('norm1', Norm, self.norm)(x) |
| | x = self.sub('mha', Attention, **kw, **ak)(x, mask, ts, training) |
| | x += skip |
| | skip = x |
| | x = self.sub('norm2', Norm, self.norm)(x) |
| | if self.glu: |
| | U = max(D, int((D * self.ffup * 2 / 3) // 32 * 32)) |
| | ff1 = self.sub('ff1', Linear, U, **kw) |
| | ff2 = self.sub('ff2', Linear, U, **kw) |
| | ff3 = self.sub('ff3', Linear, D, **kw, outscale=self.outscale) |
| | x = ff3(act(self.act)(ff1(x)) * ff2(x)) |
| | else: |
| | ff1 = self.sub('ff1', Linear, D * self.ffup, **kw) |
| | ff2 = self.sub('ff2', Linear, D, **kw, outscale=self.outscale) |
| | x = ff2(act(self.act)(ff1(x))) |
| | x += skip |
| | x = self.sub('outnorm', Norm, self.norm)(x) |
| | return x |
| |
|
| |
|
| | class GRU(nj.Module): |
| |
|
| | units: int = 1024 |
| | bias: bool = True |
| | winit: str | Callable = Initializer('trunc_normal') |
| | binit: str | Callable = Initializer('zeros') |
| | norm: str = 'rms' |
| | update_bias: float = -1.0 |
| |
|
| | def initial(self, batch_size): |
| | return jnp.zeros((batch_size, self.units), COMPUTE_DTYPE) |
| |
|
| | def __call__(self, carry, inputs, resets, single=False): |
| | assert carry.dtype == COMPUTE_DTYPE, carry.dtype |
| | assert inputs.dtype == COMPUTE_DTYPE, inputs.dtype |
| | assert resets.dtype == bool, resets.dtype |
| | if single: |
| | return self.step(carry, inputs, resets) |
| | carry, outputs = nj.scan( |
| | lambda carry, args: self.step(carry, *args), |
| | carry, (inputs, resets), axis=1) |
| | return carry, outputs |
| |
|
| | def step(self, carry, inp, reset): |
| | |
| | |
| | kw = dict(bias=self.bias, winit=self.winit, binit=self.binit) |
| | carry = mask(carry, ~reset) |
| | x = jnp.concatenate([carry, inp], -1) |
| | x = self.sub('norm', Norm, self.norm)(x) |
| | x = self.sub('linear', Linear, 3 * self.units, **kw)(x) |
| | res, cand, update = jnp.split(x, 3, -1) |
| | cand = jnp.tanh(jax.nn.sigmoid(res) * cand) |
| | update = jax.nn.sigmoid(update + self.update_bias) |
| | carry = output = update * cand + (1 - update) * carry |
| | return carry, output |
| |
|
| |
|