|
|
|
|
|
|
|
|
from __future__ import absolute_import |
|
|
from __future__ import division |
|
|
from __future__ import print_function |
|
|
|
|
|
import six |
|
|
import tensorflow as tf |
|
|
import tensorflow.contrib as tc |
|
|
|
|
|
from tensorflow.python.training import device_setter |
|
|
from tensorflow.python.framework import device as pydev |
|
|
from tensorflow.core.framework import node_def_pb2 |
|
|
|
|
|
from utils import dtype |
|
|
|
|
|
|
|
|
def local_device_setter(num_devices=1, |
|
|
ps_device_type='cpu', |
|
|
worker_device='/cpu:0', |
|
|
ps_ops=None, |
|
|
ps_strategy=None): |
|
|
if ps_ops is None: |
|
|
ps_ops = ['Variable', 'VariableV2', 'VarHandleOp'] |
|
|
|
|
|
if ps_strategy is None: |
|
|
ps_strategy = device_setter._RoundRobinStrategy(num_devices) |
|
|
if not six.callable(ps_strategy): |
|
|
raise TypeError("ps_strategy must be callable") |
|
|
|
|
|
def _local_device_chooser(op): |
|
|
current_device = pydev.DeviceSpec.from_string(op.device or "") |
|
|
|
|
|
node_def = op if isinstance(op, node_def_pb2.NodeDef) else op.node_def |
|
|
if node_def.op in ps_ops: |
|
|
ps_device_spec = pydev.DeviceSpec.from_string( |
|
|
'/{}:{}'.format(ps_device_type, ps_strategy(op))) |
|
|
|
|
|
ps_device_spec.merge_from(current_device) |
|
|
return ps_device_spec.to_string() |
|
|
else: |
|
|
worker_device_spec = pydev.DeviceSpec.from_string(worker_device or "") |
|
|
worker_device_spec.merge_from(current_device) |
|
|
return worker_device_spec.to_string() |
|
|
|
|
|
return _local_device_chooser |
|
|
|
|
|
|
|
|
def _maybe_repeat(x, n): |
|
|
if isinstance(x, list): |
|
|
assert len(x) == n |
|
|
return x |
|
|
else: |
|
|
return [x] * n |
|
|
|
|
|
|
|
|
def _reshape_output(outputs): |
|
|
|
|
|
|
|
|
if isinstance(outputs[0], (tuple, list)): |
|
|
outputs = list(zip(*outputs)) |
|
|
outputs = tuple([list(o) for o in outputs]) |
|
|
else: |
|
|
if not isinstance(outputs[0], dict): |
|
|
return outputs |
|
|
|
|
|
assert isinstance(outputs[0], dict), \ |
|
|
'invalid data type %s' % type(outputs[0]) |
|
|
|
|
|
combine_outputs = {} |
|
|
for key in outputs[0]: |
|
|
combine_outputs[key] = [o[key] for o in outputs] |
|
|
outputs = combine_outputs |
|
|
|
|
|
return outputs |
|
|
|
|
|
|
|
|
|
|
|
def data_parallelism(device_type, num_devices, fn, *args, **kwargs): |
|
|
|
|
|
if args: |
|
|
new_args = [_maybe_repeat(arg, num_devices) for arg in args] |
|
|
|
|
|
new_args = [list(x) for x in zip(*new_args)] |
|
|
else: |
|
|
new_args = [[] for _ in range(num_devices)] |
|
|
|
|
|
new_kwargs = [{} for _ in range(num_devices)] |
|
|
|
|
|
for k, v in kwargs.items(): |
|
|
vals = _maybe_repeat(v, num_devices) |
|
|
|
|
|
for i in range(num_devices): |
|
|
new_kwargs[i][k] = vals[i] |
|
|
|
|
|
fns = _maybe_repeat(fn, num_devices) |
|
|
|
|
|
|
|
|
outputs = [] |
|
|
for i in range(num_devices): |
|
|
worker = "/{}:{}".format(device_type, i) |
|
|
if device_type == 'cpu': |
|
|
_device_setter = local_device_setter(worker_device=worker) |
|
|
else: |
|
|
_device_setter = local_device_setter( |
|
|
ps_device_type='gpu', |
|
|
worker_device=worker, |
|
|
ps_strategy=tc.training.GreedyLoadBalancingStrategy( |
|
|
num_devices, tc.training.byte_size_load_fn) |
|
|
) |
|
|
|
|
|
with tf.variable_scope(tf.get_variable_scope(), reuse=bool(i != 0), |
|
|
dtype=tf.as_dtype(dtype.floatx())): |
|
|
with tf.name_scope("tower_%d" % i): |
|
|
with tf.device(_device_setter): |
|
|
outputs.append(fns[i](*new_args[i], **new_kwargs[i])) |
|
|
|
|
|
return _reshape_output(outputs) |
|
|
|
|
|
|
|
|
def parallel_model(model_fn, features, devices, use_cpu=False): |
|
|
device_type = 'gpu' |
|
|
num_devices = len(devices) |
|
|
|
|
|
if use_cpu: |
|
|
device_type = 'cpu' |
|
|
num_devices = 1 |
|
|
|
|
|
outputs = data_parallelism(device_type, num_devices, model_fn, features) |
|
|
|
|
|
return outputs |
|
|
|
|
|
|
|
|
def average_gradients(tower_grads, mask=None): |
|
|
"""Modified from Bilm""" |
|
|
|
|
|
|
|
|
if len(tower_grads) == 1: |
|
|
return tower_grads[0] |
|
|
|
|
|
|
|
|
def _deduplicate_indexed_slices(values, indices): |
|
|
"""Sums `values` associated with any non-unique `indices`.""" |
|
|
unique_indices, new_index_positions = tf.unique(indices) |
|
|
summed_values = tf.unsorted_segment_sum( |
|
|
values, new_index_positions, |
|
|
tf.shape(unique_indices)[0]) |
|
|
return summed_values, unique_indices |
|
|
|
|
|
average_grads = [] |
|
|
for grad_and_vars in zip(*tower_grads): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
g0, v0 = grad_and_vars[0] |
|
|
|
|
|
if g0 is None: |
|
|
|
|
|
tf.logging.warn("{} has no gradient".format(v0.name)) |
|
|
average_grads.append((g0, v0)) |
|
|
continue |
|
|
|
|
|
if isinstance(g0, tf.IndexedSlices): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
indices = [] |
|
|
values = [] |
|
|
for g, v in grad_and_vars: |
|
|
indices.append(g.indices) |
|
|
values.append(g.values) |
|
|
all_indices = tf.concat(indices, 0) |
|
|
if mask is None: |
|
|
avg_values = tf.concat(values, 0) / len(grad_and_vars) |
|
|
else: |
|
|
avg_values = tf.concat(values, 0) / tf.reduce_sum(mask) |
|
|
|
|
|
av, ai = _deduplicate_indexed_slices(avg_values, all_indices) |
|
|
grad = tf.IndexedSlices(av, ai, dense_shape=g0.dense_shape) |
|
|
else: |
|
|
|
|
|
grads = [] |
|
|
for g, v in grad_and_vars: |
|
|
|
|
|
expanded_g = tf.expand_dims(g, 0) |
|
|
|
|
|
grads.append(expanded_g) |
|
|
|
|
|
|
|
|
grad = tf.concat(grads, 0) |
|
|
if mask is not None: |
|
|
grad = tf.boolean_mask( |
|
|
grad, tf.cast(mask, tf.bool), axis=0) |
|
|
grad = tf.reduce_mean(grad, 0) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
v = grad_and_vars[0][1] |
|
|
grad_and_var = (grad, v) |
|
|
|
|
|
average_grads.append(grad_and_var) |
|
|
|
|
|
assert len(average_grads) == len(list(zip(*tower_grads))) |
|
|
|
|
|
return average_grads |
|
|
|