|
|
import threading |
|
|
import re |
|
|
from collections import Counter |
|
|
|
|
|
import jax |
|
|
from jax.sharding import PartitionSpec as P |
|
|
import ninjax as nj |
|
|
|
|
|
from . import nets as nn |
|
|
|
|
|
|
|
|
LOCK = threading.Lock() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
TRACER_SHARDINGS = {} |
|
|
|
|
|
|
|
|
def init( |
|
|
fn, mesh, arg_shardings, |
|
|
param_partition_rules=(), |
|
|
act_partition_rules=(), |
|
|
static_argnums=(), |
|
|
dummy_inputs=(), |
|
|
print_partition=False, |
|
|
): |
|
|
|
|
|
def init(fun, **jit_kwargs): |
|
|
if not getattr(fun, '_is_pure', False): |
|
|
fun = nj.pure(fun) |
|
|
def wrapper(*args, **kwargs): |
|
|
state, out = fun(*args, create=True, modify=True, ignore=True, **kwargs) |
|
|
del out |
|
|
return state, () |
|
|
return wrapper |
|
|
fn = init(fn) |
|
|
|
|
|
def fn(*args, inner=fn): |
|
|
params, seed, *args = args |
|
|
old = nn.LAYER_CALLBACK |
|
|
nn.LAYER_CALLBACK = create_layer_callback(mesh, act_partition_rules) |
|
|
params, _ = inner(params, *args, seed=seed) |
|
|
nn.LAYER_CALLBACK = old |
|
|
return params |
|
|
|
|
|
fn = jax.jit(fn, static_argnums=static_argnums) |
|
|
|
|
|
params_shapes = fn.eval_shape(*dummy_inputs) |
|
|
params_sharding, grouping = resolve_rules( |
|
|
params_shapes, param_partition_rules, mesh) |
|
|
if print_partition: |
|
|
print_grouping(grouping) |
|
|
|
|
|
fn = jax.jit(fn, arg_shardings, params_sharding, static_argnums, None) |
|
|
params = fn(*dummy_inputs) |
|
|
|
|
|
return params, params_sharding |
|
|
|
|
|
|
|
|
def apply( |
|
|
fn, mesh, in_shardings, out_shardings, |
|
|
partition_rules=(), |
|
|
static_argnums=(), |
|
|
single_output=False, |
|
|
return_params=False, |
|
|
donate_params=False, |
|
|
|
|
|
split_rng=True, |
|
|
use_shardmap=False, |
|
|
first_outnums=(), |
|
|
): |
|
|
|
|
|
if single_output: |
|
|
assert len(out_shardings) == 1 |
|
|
|
|
|
def fn(*args, inner=fn): |
|
|
if donate_params: |
|
|
donated, allocated, seed, *args = args |
|
|
params = {**donated, **allocated} |
|
|
else: |
|
|
params, seed, *args = args |
|
|
if use_shardmap and len(mesh.devices) > 1 and split_rng: |
|
|
seed = jax.random.fold_in(seed, jax.lax.axis_index('d')) |
|
|
params, outs = inner(params, *args, seed=seed) |
|
|
outs = (outs,) if single_output else outs |
|
|
assert isinstance(outs, tuple) |
|
|
return (params, *outs) if return_params else outs |
|
|
|
|
|
if use_shardmap and len(mesh.devices) > 1: |
|
|
|
|
|
def fn(*args, inner=fn): |
|
|
outs = list(inner(*args)) |
|
|
for i in first_outnums: |
|
|
outs[i] = jax.tree.map(lambda x: x[None], outs[i]) |
|
|
return tuple(outs) |
|
|
|
|
|
from jax.experimental.shard_map import shard_map |
|
|
ispecs = list(jax.tree.map(lambda s: s.spec, in_shardings)) |
|
|
for i in sorted(static_argnums): |
|
|
ispecs.insert(i, None) |
|
|
ispecs = tuple(ispecs) |
|
|
ospecs = jax.tree.map(lambda s: s.spec, out_shardings) |
|
|
fn = shard_map(fn, mesh, ispecs, ospecs, check_rep=False) |
|
|
|
|
|
def fn(*args, inner=fn): |
|
|
outs = list(inner(*args)) |
|
|
for i in first_outnums: |
|
|
outs[i] = jax.tree.map(lambda x: x[0], outs[i]) |
|
|
return tuple(outs) |
|
|
|
|
|
if single_output: |
|
|
def fn(*args, inner=fn): |
|
|
outs = inner(*args) |
|
|
assert len(outs) == 1 |
|
|
return outs[0] |
|
|
|
|
|
if single_output: |
|
|
out_shardings = out_shardings[0] |
|
|
donate = [0] if donate_params else [] |
|
|
|
|
|
if not use_shardmap: |
|
|
def fn(*args, inner=fn): |
|
|
with LOCK: |
|
|
old = nn.LAYER_CALLBACK |
|
|
nn.LAYER_CALLBACK = create_layer_callback(mesh, partition_rules) |
|
|
outs = inner(*args) |
|
|
nn.LAYER_CALLBACK = old |
|
|
return outs |
|
|
|
|
|
fn = jax.jit(fn, in_shardings, out_shardings, static_argnums, None, donate) |
|
|
|
|
|
return fn |
|
|
|
|
|
|
|
|
def create_layer_callback(mesh, partition_rules): |
|
|
def layer_callback(y, name): |
|
|
name = f'{nj.ninjax.SCOPE}/{name}' |
|
|
for rule, spec in partition_rules: |
|
|
if re.search(rule, name): |
|
|
sharding = jax.sharding.NamedSharding(mesh, spec) |
|
|
def apply(y): |
|
|
y = jax.lax.with_sharding_constraint(y, sharding) |
|
|
if not hasattr(type(y), 'tracer_shardings'): |
|
|
type(y).tracer_sharding = property( |
|
|
lambda self: TRACER_SHARDINGS[id(self)]) |
|
|
TRACER_SHARDINGS[id(y)] = sharding |
|
|
return y |
|
|
return jax.tree.map(apply, y) |
|
|
else: |
|
|
raise Exception(f'No matching rule found for activation key: {name}') |
|
|
return layer_callback |
|
|
|
|
|
|
|
|
def resolve_rules(params, partition_rules, mesh): |
|
|
if len(partition_rules) == 0: |
|
|
partition_rules = [('.*', P())] |
|
|
params_spec, grouping = dict(), dict() |
|
|
for k in params.keys(): |
|
|
for rule, spec in partition_rules: |
|
|
if re.search(rule, k): |
|
|
params_spec[k] = spec |
|
|
if rule not in grouping: |
|
|
grouping[rule] = [] |
|
|
grouping[rule].append(k) |
|
|
break |
|
|
else: |
|
|
raise Exception(f'No matching rule found for param key: {k}') |
|
|
assert set(params.keys()) == set(params_spec.keys()) |
|
|
sharding = jax.tree.map( |
|
|
lambda spec: jax.sharding.NamedSharding(mesh, spec), params_spec) |
|
|
return sharding, grouping |
|
|
|
|
|
|
|
|
def print_grouping(grouping): |
|
|
for rule, ps in grouping.items(): |
|
|
if len(ps) == 0: |
|
|
continue |
|
|
print(f'Partition rule "{rule}" matches {len(ps)} param tensors') |
|
|
ks = ['/'.join(p.split('/')[-2:]) for p in ps] |
|
|
ks = Counter(ks) |
|
|
ks = ks.most_common(len(ks)) |
|
|
ks = [f'- .../{k}: {v}' for k, v in ks] |
|
|
print('\n'.join(ks)) |
|
|
|