# coding: utf-8 from __future__ import absolute_import from __future__ import division from __future__ import print_function import os import time import collections import numpy as np import tensorflow as tf from utils import dtype def batch_indexer(datasize, batch_size): """Just divide the datasize into batched size""" dataindex = np.arange(datasize).tolist() batchindex = [] for i in range(datasize // batch_size): batchindex.append(dataindex[i * batch_size: (i + 1) * batch_size]) if datasize % batch_size > 0: batchindex.append(dataindex[-(datasize % batch_size):]) return batchindex def token_indexer(dataset, token_size): """Divide the dataset into token-based batch""" # assume dataset format: [(len1, len2, ..., lenN)] dataindex = np.arange(len(dataset)).tolist() batchindex = [] _batcher = [0.] * len(dataset[0]) _counter = 0 i = 0 while True: if i >= len(dataset): break # attempt put this datapoint into batch _batcher = [max(max_l, l) for max_l, l in zip(_batcher, dataset[i])] _counter += 1 for l in _batcher: if _counter * l >= token_size: # when an extreme instance occur, handle it by making a 1-size batch if _counter > 1: batchindex.append(dataindex[i-_counter+1: i]) i -= 1 else: batchindex.append(dataindex[i: i+1]) _counter = 0 _batcher = [0.] * len(dataset[0]) break i += 1 _counter = sum([len(slice) for slice in batchindex]) if _counter != len(dataset): batchindex.append(dataindex[_counter:]) return batchindex def mask_scale(value, mask, scale=None): """Prepared for masked softmax""" if scale is None: scale = dtype.inf() return value + (1. - mask) * (-scale) def valid_apply_dropout(x, dropout): """To check whether the dropout value is valid, apply if valid""" if dropout is not None and 0. <= dropout <= 1.: return tf.nn.dropout(x, 1. - dropout) return x def layer_dropout(dropped, no_dropped, dropout_rate): """Layer Dropout""" pred = tf.random_uniform([]) < dropout_rate return tf.cond(pred, lambda: dropped, lambda: no_dropped) def label_smooth(labels, vocab_size, factor=0.1): """Smooth the gold label distribution""" if 0. < factor < 1.: n = tf.cast(vocab_size - 1, tf.float32) p = 1. - factor q = factor / n t = tf.one_hot(tf.cast(tf.reshape(labels, [-1]), tf.int32), depth=vocab_size, on_value=p, off_value=q) normalizing = -(p * tf.log(p) + n * q * tf.log(q + 1e-20)) else: t = tf.one_hot(tf.cast(tf.reshape(labels, [-1]), tf.int32), depth=vocab_size) normalizing = 0. return t, normalizing def closing_dropout(params): """Removing all dropouts""" for k, v in params.values().items(): if 'dropout' in k: setattr(params, k, 0.0) # consider closing label smoothing if 'label_smoothing' in k: setattr(params, k, 0.0) return params def dict_update(d, u): """Recursive update dictionary""" for k, v in u.items(): if isinstance(v, collections.Mapping): d[k] = dict_update(d.get(k, {}), v) else: d[k] = v return d def shape_list(x): # Copied from Tensor2Tensor """Return list of dims, statically where possible.""" x = tf.convert_to_tensor(x) # If unknown rank, return dynamic shape if x.get_shape().dims is None: return tf.shape(x) static = x.get_shape().as_list() shape = tf.shape(x) ret = [] for i in range(len(static)): dim = static[i] if dim is None: dim = shape[i] ret.append(dim) return ret def get_shape_invariants(tensor): # Copied from Tensor2Tensor """Returns the shape of the tensor but sets middle dims to None.""" shape = tensor.shape.as_list() for i in range(1, len(shape) - 1): shape[i] = None return tf.TensorShape(shape) def merge_neighbor_dims(x, axis=0): """Merge neighbor dimension of x, start by axis""" if len(x.get_shape().as_list()) < axis + 2: return x shape = shape_list(x) shape[axis] *= shape[axis+1] shape.pop(axis+1) return tf.reshape(x, shape) def unmerge_neighbor_dims(x, depth, axis=0): """Inverse of merge_neighbor_dims, axis by depth""" if len(x.get_shape().as_list()) < axis + 1: return x shape = shape_list(x) width = shape[axis] // depth new_shape = shape[:axis] + [depth, width] + shape[axis+1:] return tf.reshape(x, new_shape) def expand_tile_dims(x, depth, axis=1): """Expand and Tile x on axis by depth""" x = tf.expand_dims(x, axis=axis) tile_dims = [1] * x.shape.ndims tile_dims[axis] = depth return tf.tile(x, tile_dims) def gumbel_noise(shape, eps=None): """Generate gumbel noise shaped by shape""" if eps is None: eps = dtype.epsilon() u = tf.random_uniform(shape, minval=0, maxval=1) return -tf.log(-tf.log(u + eps) + eps) def log_prob_from_logits(logits): """Probability from un-nomalized logits""" return logits - tf.reduce_logsumexp(logits, axis=-1, keepdims=True) def batch_coordinates(batch_size, beam_size): """Batch coordinate indices under beam_size""" batch_pos = tf.range(batch_size * beam_size) // beam_size batch_pos = tf.reshape(batch_pos, [batch_size, beam_size]) return batch_pos def variable_printer(): """Print parameters""" all_weights = {v.name: v for v in tf.trainable_variables()} total_size = 0 for v_name in sorted(list(all_weights)): v = all_weights[v_name] tf.logging.info("%s\tshape %s", v.name[:-2].ljust(80), str(v.shape).ljust(20)) v_size = np.prod(np.array(v.shape.as_list())).tolist() total_size += v_size tf.logging.info("Total trainable variables size: %d", total_size) def uniform_splits(total_size, num_shards): """Split the total_size into uniform num_shards lists""" size_per_shards = total_size // num_shards splits = [size_per_shards] * (num_shards - 1) + \ [total_size - (num_shards - 1) * size_per_shards] return splits def fetch_valid_ref_files(path): """Extracting valid reference files according to MT convention""" path = os.path.abspath(path) if tf.gfile.Exists(path): return [path] if not tf.gfile.Exists(path + ".ref0"): tf.logging.warn("Invalid Reference Format {}".format(path)) return None num = 0 files = [] while True: file_path = path + ".ref%s" % num if tf.gfile.Exists(file_path): files.append(file_path) else: break num += 1 return files def get_session(gpus): """Config session with GPUS""" sess_config = tf.ConfigProto(allow_soft_placement=True) sess_config.gpu_options.allow_growth = True if len(gpus) > 0: device_str = ",".join([str(i) for i in gpus]) sess_config.gpu_options.visible_device_list = device_str sess = tf.Session(config=sess_config) return sess def flatten_list(values): """Flatten a list""" return [v for value in values for v in value] def time_str(t=None): """String format of the time long data""" if t is None: t = time.time() ts = time.strftime("[%Y-%m-%d %H:%M:%S]", time.localtime(t)) return ts