repo stringlengths 7 55 | path stringlengths 4 223 | func_name stringlengths 1 134 | original_string stringlengths 75 104k | language stringclasses 1
value | code stringlengths 75 104k | code_tokens listlengths 19 28.4k | docstring stringlengths 1 46.9k | docstring_tokens listlengths 1 1.97k | sha stringlengths 40 40 | url stringlengths 87 315 | partition stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | bit_to_int | def bit_to_int(x_bit, num_bits, base=2):
"""Turn x_bit representing numbers bitwise (lower-endian) to int tensor.
Args:
x_bit: Tensor containing numbers in a particular base to be converted to
int.
num_bits: Number of bits in the representation.
base: Base of the representation.
Returns:
I... | python | def bit_to_int(x_bit, num_bits, base=2):
"""Turn x_bit representing numbers bitwise (lower-endian) to int tensor.
Args:
x_bit: Tensor containing numbers in a particular base to be converted to
int.
num_bits: Number of bits in the representation.
base: Base of the representation.
Returns:
I... | [
"def",
"bit_to_int",
"(",
"x_bit",
",",
"num_bits",
",",
"base",
"=",
"2",
")",
":",
"x_l",
"=",
"tf",
".",
"stop_gradient",
"(",
"tf",
".",
"to_int32",
"(",
"tf",
".",
"reshape",
"(",
"x_bit",
",",
"[",
"-",
"1",
",",
"num_bits",
"]",
")",
")",
... | Turn x_bit representing numbers bitwise (lower-endian) to int tensor.
Args:
x_bit: Tensor containing numbers in a particular base to be converted to
int.
num_bits: Number of bits in the representation.
base: Base of the representation.
Returns:
Integer representation of this number. | [
"Turn",
"x_bit",
"representing",
"numbers",
"bitwise",
"(",
"lower",
"-",
"endian",
")",
"to",
"int",
"tensor",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L225-L241 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | int_to_bit_embed | def int_to_bit_embed(x_int, num_bits, embedding_size, base=2):
"""Turn x_int into a bitwise (lower-endian) tensor and embed densly."""
shape = common_layers.shape_list(x_int)
inputs = int_to_bit(x_int, num_bits, base=base)
inputs = tf.reshape(inputs, shape[:-1] + [shape[-1] * 8])
inputs = 2.0 * tf.to_float(in... | python | def int_to_bit_embed(x_int, num_bits, embedding_size, base=2):
"""Turn x_int into a bitwise (lower-endian) tensor and embed densly."""
shape = common_layers.shape_list(x_int)
inputs = int_to_bit(x_int, num_bits, base=base)
inputs = tf.reshape(inputs, shape[:-1] + [shape[-1] * 8])
inputs = 2.0 * tf.to_float(in... | [
"def",
"int_to_bit_embed",
"(",
"x_int",
",",
"num_bits",
",",
"embedding_size",
",",
"base",
"=",
"2",
")",
":",
"shape",
"=",
"common_layers",
".",
"shape_list",
"(",
"x_int",
")",
"inputs",
"=",
"int_to_bit",
"(",
"x_int",
",",
"num_bits",
",",
"base",
... | Turn x_int into a bitwise (lower-endian) tensor and embed densly. | [
"Turn",
"x_int",
"into",
"a",
"bitwise",
"(",
"lower",
"-",
"endian",
")",
"tensor",
"and",
"embed",
"densly",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L263-L269 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | embed | def embed(x,
hidden_size,
z_size,
filter_size,
bottleneck_kind="dvq",
soft_em=False,
num_blocks=2,
num_residuals=1,
block_v_size=None,
means=None,
name=None):
"""Embedding function that takes discrete latent and return... | python | def embed(x,
hidden_size,
z_size,
filter_size,
bottleneck_kind="dvq",
soft_em=False,
num_blocks=2,
num_residuals=1,
block_v_size=None,
means=None,
name=None):
"""Embedding function that takes discrete latent and return... | [
"def",
"embed",
"(",
"x",
",",
"hidden_size",
",",
"z_size",
",",
"filter_size",
",",
"bottleneck_kind",
"=",
"\"dvq\"",
",",
"soft_em",
"=",
"False",
",",
"num_blocks",
"=",
"2",
",",
"num_residuals",
"=",
"1",
",",
"block_v_size",
"=",
"None",
",",
"me... | Embedding function that takes discrete latent and returns embedding.
Args:
x: Input to the discretization bottleneck.
hidden_size: Dimension of the latent state.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
filter_size: Dimension to project embedding by. Used only if bottle... | [
"Embedding",
"function",
"that",
"takes",
"discrete",
"latent",
"and",
"returns",
"embedding",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L272-L357 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vae | def vae(x, z_size, name=None):
"""Simple variational autoencoder without discretization.
Args:
x: Input to the discretization bottleneck.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
name: Name for the bottleneck scope.
Returns:
Embedding function, latent, loss, mu and... | python | def vae(x, z_size, name=None):
"""Simple variational autoencoder without discretization.
Args:
x: Input to the discretization bottleneck.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
name: Name for the bottleneck scope.
Returns:
Embedding function, latent, loss, mu and... | [
"def",
"vae",
"(",
"x",
",",
"z_size",
",",
"name",
"=",
"None",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"name",
",",
"default_name",
"=",
"\"vae\"",
")",
":",
"mu",
"=",
"tf",
".",
"layers",
".",
"dense",
"(",
"x",
",",
"z_size",
","... | Simple variational autoencoder without discretization.
Args:
x: Input to the discretization bottleneck.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
name: Name for the bottleneck scope.
Returns:
Embedding function, latent, loss, mu and log_simga. | [
"Simple",
"variational",
"autoencoder",
"without",
"discretization",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L360-L381 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | gumbel_sample | def gumbel_sample(shape):
"""Sample from the Gumbel distribution, protect from overflows.
Args:
shape: Shape of Gumbel samples.
Returns:
Noise drawn from Gumbel distribution.
"""
uniform_samples = tf.random_uniform(shape, minval=0.00001, maxval=0.99998)
return -tf.log(-tf.log(uniform_samples)) | python | def gumbel_sample(shape):
"""Sample from the Gumbel distribution, protect from overflows.
Args:
shape: Shape of Gumbel samples.
Returns:
Noise drawn from Gumbel distribution.
"""
uniform_samples = tf.random_uniform(shape, minval=0.00001, maxval=0.99998)
return -tf.log(-tf.log(uniform_samples)) | [
"def",
"gumbel_sample",
"(",
"shape",
")",
":",
"uniform_samples",
"=",
"tf",
".",
"random_uniform",
"(",
"shape",
",",
"minval",
"=",
"0.00001",
",",
"maxval",
"=",
"0.99998",
")",
"return",
"-",
"tf",
".",
"log",
"(",
"-",
"tf",
".",
"log",
"(",
"u... | Sample from the Gumbel distribution, protect from overflows.
Args:
shape: Shape of Gumbel samples.
Returns:
Noise drawn from Gumbel distribution. | [
"Sample",
"from",
"the",
"Gumbel",
"distribution",
"protect",
"from",
"overflows",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L402-L412 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | gumbel_softmax | def gumbel_softmax(x,
z_size,
mode,
softmax_k=0,
temperature_warmup_steps=150000,
summary=True,
name=None):
"""Gumbel softmax discretization bottleneck.
Args:
x: Input to the discretization bottlen... | python | def gumbel_softmax(x,
z_size,
mode,
softmax_k=0,
temperature_warmup_steps=150000,
summary=True,
name=None):
"""Gumbel softmax discretization bottleneck.
Args:
x: Input to the discretization bottlen... | [
"def",
"gumbel_softmax",
"(",
"x",
",",
"z_size",
",",
"mode",
",",
"softmax_k",
"=",
"0",
",",
"temperature_warmup_steps",
"=",
"150000",
",",
"summary",
"=",
"True",
",",
"name",
"=",
"None",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"name",
... | Gumbel softmax discretization bottleneck.
Args:
x: Input to the discretization bottleneck.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
mode: tf.estimator.ModeKeys.
softmax_k: If > 0 then do top-k softmax.
temperature_warmup_steps: Number of steps it takes to decay temp... | [
"Gumbel",
"softmax",
"discretization",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L415-L475 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | discrete_bottleneck | def discrete_bottleneck(inputs,
hidden_size,
z_size,
filter_size,
mode=None,
bottleneck_kind="dvq",
num_blocks=2,
num_residuals=1,
... | python | def discrete_bottleneck(inputs,
hidden_size,
z_size,
filter_size,
mode=None,
bottleneck_kind="dvq",
num_blocks=2,
num_residuals=1,
... | [
"def",
"discrete_bottleneck",
"(",
"inputs",
",",
"hidden_size",
",",
"z_size",
",",
"filter_size",
",",
"mode",
"=",
"None",
",",
"bottleneck_kind",
"=",
"\"dvq\"",
",",
"num_blocks",
"=",
"2",
",",
"num_residuals",
"=",
"1",
",",
"reshape_method",
"=",
"\"... | Discretization bottleneck.
Args:
inputs: Input to the bottleneck, a Tensor of shape [..., channels].
hidden_size: Dimension of the dense output.
z_size: Number of bits, where discrete codes range from 1 to 2**z_size.
filter_size: Filter size in the embedding function.
mode: tf.estimator.ModeKeys.... | [
"Discretization",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L478-L788 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | predict_bits_with_lstm | def predict_bits_with_lstm(prediction_source, state_size, total_num_bits,
target_bits=None, extra_inputs=None,
bits_at_once=8, temperature=1.0, dropout=0.1):
"""Predict a sequence of bits (a latent) with LSTM, both training and infer.
Given a tensor on which th... | python | def predict_bits_with_lstm(prediction_source, state_size, total_num_bits,
target_bits=None, extra_inputs=None,
bits_at_once=8, temperature=1.0, dropout=0.1):
"""Predict a sequence of bits (a latent) with LSTM, both training and infer.
Given a tensor on which th... | [
"def",
"predict_bits_with_lstm",
"(",
"prediction_source",
",",
"state_size",
",",
"total_num_bits",
",",
"target_bits",
"=",
"None",
",",
"extra_inputs",
"=",
"None",
",",
"bits_at_once",
"=",
"8",
",",
"temperature",
"=",
"1.0",
",",
"dropout",
"=",
"0.1",
"... | Predict a sequence of bits (a latent) with LSTM, both training and infer.
Given a tensor on which the predictions are based (prediction_source), we use
a single-layer LSTM with state of size state_size to predict total_num_bits,
which we predict in groups of size bits_at_once. During training, we use
target_bi... | [
"Predict",
"a",
"sequence",
"of",
"bits",
"(",
"a",
"latent",
")",
"with",
"LSTM",
"both",
"training",
"and",
"infer",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L791-L876 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | get_vq_codebook | def get_vq_codebook(codebook_size, hidden_size):
"""Get lookup table for VQ bottleneck."""
with tf.variable_scope("vq", reuse=tf.AUTO_REUSE):
means = tf.get_variable(
name="means",
shape=[codebook_size, hidden_size],
initializer=tf.uniform_unit_scaling_initializer())
ema_count = tf.... | python | def get_vq_codebook(codebook_size, hidden_size):
"""Get lookup table for VQ bottleneck."""
with tf.variable_scope("vq", reuse=tf.AUTO_REUSE):
means = tf.get_variable(
name="means",
shape=[codebook_size, hidden_size],
initializer=tf.uniform_unit_scaling_initializer())
ema_count = tf.... | [
"def",
"get_vq_codebook",
"(",
"codebook_size",
",",
"hidden_size",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"\"vq\"",
",",
"reuse",
"=",
"tf",
".",
"AUTO_REUSE",
")",
":",
"means",
"=",
"tf",
".",
"get_variable",
"(",
"name",
"=",
"\"means\"",
... | Get lookup table for VQ bottleneck. | [
"Get",
"lookup",
"table",
"for",
"VQ",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L885-L905 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vq_nearest_neighbor | def vq_nearest_neighbor(x, means,
soft_em=False, num_samples=10, temperature=None):
"""Find the nearest element in means to elements in x."""
bottleneck_size = common_layers.shape_list(means)[0]
x_norm_sq = tf.reduce_sum(tf.square(x), axis=-1, keepdims=True)
means_norm_sq = tf.reduce_sum... | python | def vq_nearest_neighbor(x, means,
soft_em=False, num_samples=10, temperature=None):
"""Find the nearest element in means to elements in x."""
bottleneck_size = common_layers.shape_list(means)[0]
x_norm_sq = tf.reduce_sum(tf.square(x), axis=-1, keepdims=True)
means_norm_sq = tf.reduce_sum... | [
"def",
"vq_nearest_neighbor",
"(",
"x",
",",
"means",
",",
"soft_em",
"=",
"False",
",",
"num_samples",
"=",
"10",
",",
"temperature",
"=",
"None",
")",
":",
"bottleneck_size",
"=",
"common_layers",
".",
"shape_list",
"(",
"means",
")",
"[",
"0",
"]",
"x... | Find the nearest element in means to elements in x. | [
"Find",
"the",
"nearest",
"element",
"in",
"means",
"to",
"elements",
"in",
"x",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L908-L934 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vq_discrete_bottleneck | def vq_discrete_bottleneck(x,
bottleneck_bits,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10):
"""Simple vector quantized discrete bot... | python | def vq_discrete_bottleneck(x,
bottleneck_bits,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10):
"""Simple vector quantized discrete bot... | [
"def",
"vq_discrete_bottleneck",
"(",
"x",
",",
"bottleneck_bits",
",",
"beta",
"=",
"0.25",
",",
"decay",
"=",
"0.999",
",",
"epsilon",
"=",
"1e-5",
",",
"soft_em",
"=",
"False",
",",
"num_samples",
"=",
"10",
")",
":",
"bottleneck_size",
"=",
"2",
"**"... | Simple vector quantized discrete bottleneck. | [
"Simple",
"vector",
"quantized",
"discrete",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L937-L954 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vq_body | def vq_body(x,
codebook_size,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10,
temperature=None,
do_update=True):
"""Discretize each x into one of codebook_size codes."""
x_shape = common_layers.shape... | python | def vq_body(x,
codebook_size,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10,
temperature=None,
do_update=True):
"""Discretize each x into one of codebook_size codes."""
x_shape = common_layers.shape... | [
"def",
"vq_body",
"(",
"x",
",",
"codebook_size",
",",
"beta",
"=",
"0.25",
",",
"decay",
"=",
"0.999",
",",
"epsilon",
"=",
"1e-5",
",",
"soft_em",
"=",
"False",
",",
"num_samples",
"=",
"10",
",",
"temperature",
"=",
"None",
",",
"do_update",
"=",
... | Discretize each x into one of codebook_size codes. | [
"Discretize",
"each",
"x",
"into",
"one",
"of",
"codebook_size",
"codes",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L957-L1004 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vq_loss | def vq_loss(x,
targets,
codebook_size,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10,
temperature=None,
do_update=True):
"""Compute the loss of large vocab tensors using a VQAE codebook.
... | python | def vq_loss(x,
targets,
codebook_size,
beta=0.25,
decay=0.999,
epsilon=1e-5,
soft_em=False,
num_samples=10,
temperature=None,
do_update=True):
"""Compute the loss of large vocab tensors using a VQAE codebook.
... | [
"def",
"vq_loss",
"(",
"x",
",",
"targets",
",",
"codebook_size",
",",
"beta",
"=",
"0.25",
",",
"decay",
"=",
"0.999",
",",
"epsilon",
"=",
"1e-5",
",",
"soft_em",
"=",
"False",
",",
"num_samples",
"=",
"10",
",",
"temperature",
"=",
"None",
",",
"d... | Compute the loss of large vocab tensors using a VQAE codebook.
Args:
x: Tensor of inputs to be quantized to nearest code
targets: Tensor of target indices to target codes
codebook_size: Size of quantization codebook
beta: scalar float for moving averages
decay: scalar float for moving averages
... | [
"Compute",
"the",
"loss",
"of",
"large",
"vocab",
"tensors",
"using",
"a",
"VQAE",
"codebook",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1007-L1071 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | vq_discrete_unbottleneck | def vq_discrete_unbottleneck(x, hidden_size):
"""Simple undiscretization from vector quantized representation."""
x_shape = common_layers.shape_list(x)
x = tf.to_float(x)
bottleneck_size = common_layers.shape_list(x)[-1]
means, _, _ = get_vq_codebook(bottleneck_size, hidden_size)
result = tf.matmul(tf.resha... | python | def vq_discrete_unbottleneck(x, hidden_size):
"""Simple undiscretization from vector quantized representation."""
x_shape = common_layers.shape_list(x)
x = tf.to_float(x)
bottleneck_size = common_layers.shape_list(x)[-1]
means, _, _ = get_vq_codebook(bottleneck_size, hidden_size)
result = tf.matmul(tf.resha... | [
"def",
"vq_discrete_unbottleneck",
"(",
"x",
",",
"hidden_size",
")",
":",
"x_shape",
"=",
"common_layers",
".",
"shape_list",
"(",
"x",
")",
"x",
"=",
"tf",
".",
"to_float",
"(",
"x",
")",
"bottleneck_size",
"=",
"common_layers",
".",
"shape_list",
"(",
"... | Simple undiscretization from vector quantized representation. | [
"Simple",
"undiscretization",
"from",
"vector",
"quantized",
"representation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1074-L1081 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | gumbel_softmax_nearest_neighbor_dvq | def gumbel_softmax_nearest_neighbor_dvq(x,
means,
block_v_size,
hard=False,
temperature_init=1.2,
num_samples=1,
... | python | def gumbel_softmax_nearest_neighbor_dvq(x,
means,
block_v_size,
hard=False,
temperature_init=1.2,
num_samples=1,
... | [
"def",
"gumbel_softmax_nearest_neighbor_dvq",
"(",
"x",
",",
"means",
",",
"block_v_size",
",",
"hard",
"=",
"False",
",",
"temperature_init",
"=",
"1.2",
",",
"num_samples",
"=",
"1",
",",
"temperature_warmup_steps",
"=",
"150000",
",",
"summary",
"=",
"True",
... | Sample from Gumbel-Softmax and compute neighbors and losses.
Args:
x: A `float`-like `Tensor` of shape [batch_size, latent_dim, num_blocks,
block_dim] containing the latent vectors to be compared to the codebook.
means: Embedding table of shape [num_blocks, block_v_size, block_dim].
block_v_size: N... | [
"Sample",
"from",
"Gumbel",
"-",
"Softmax",
"and",
"compute",
"neighbors",
"and",
"losses",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1084-L1251 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | gumbel_softmax_discrete_bottleneck | def gumbel_softmax_discrete_bottleneck(x,
bottleneck_bits,
beta=0.25,
decay=0.999,
epsilon=1e-5,
temperature_warmup_steps=150... | python | def gumbel_softmax_discrete_bottleneck(x,
bottleneck_bits,
beta=0.25,
decay=0.999,
epsilon=1e-5,
temperature_warmup_steps=150... | [
"def",
"gumbel_softmax_discrete_bottleneck",
"(",
"x",
",",
"bottleneck_bits",
",",
"beta",
"=",
"0.25",
",",
"decay",
"=",
"0.999",
",",
"epsilon",
"=",
"1e-5",
",",
"temperature_warmup_steps",
"=",
"150000",
",",
"hard",
"=",
"False",
",",
"summary",
"=",
... | VQ-VAE using Gumbel-Softmax.
Different from `gumbel_softmax()` function as
this function calculates the KL by using the discrete entropy
instead of taking the argmax, and it also uses an exponential moving average
to update the codebook while the `gumbel_softmax()` function includes no
codebook update.
Ar... | [
"VQ",
"-",
"VAE",
"using",
"Gumbel",
"-",
"Softmax",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1254-L1371 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | tanh_discrete_bottleneck | def tanh_discrete_bottleneck(x, bottleneck_bits, bottleneck_noise,
discretize_warmup_steps, mode):
"""Simple discretization through tanh, flip bottleneck_noise many bits."""
x = tf.layers.dense(x, bottleneck_bits, name="tanh_discrete_bottleneck")
d0 = tf.stop_gradient(2.0 * tf.to_floa... | python | def tanh_discrete_bottleneck(x, bottleneck_bits, bottleneck_noise,
discretize_warmup_steps, mode):
"""Simple discretization through tanh, flip bottleneck_noise many bits."""
x = tf.layers.dense(x, bottleneck_bits, name="tanh_discrete_bottleneck")
d0 = tf.stop_gradient(2.0 * tf.to_floa... | [
"def",
"tanh_discrete_bottleneck",
"(",
"x",
",",
"bottleneck_bits",
",",
"bottleneck_noise",
",",
"discretize_warmup_steps",
",",
"mode",
")",
":",
"x",
"=",
"tf",
".",
"layers",
".",
"dense",
"(",
"x",
",",
"bottleneck_bits",
",",
"name",
"=",
"\"tanh_discre... | Simple discretization through tanh, flip bottleneck_noise many bits. | [
"Simple",
"discretization",
"through",
"tanh",
"flip",
"bottleneck_noise",
"many",
"bits",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1374-L1390 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | tanh_discrete_unbottleneck | def tanh_discrete_unbottleneck(x, hidden_size):
"""Simple un-discretization from tanh."""
x = tf.layers.dense(x, hidden_size, name="tanh_discrete_unbottleneck")
return x | python | def tanh_discrete_unbottleneck(x, hidden_size):
"""Simple un-discretization from tanh."""
x = tf.layers.dense(x, hidden_size, name="tanh_discrete_unbottleneck")
return x | [
"def",
"tanh_discrete_unbottleneck",
"(",
"x",
",",
"hidden_size",
")",
":",
"x",
"=",
"tf",
".",
"layers",
".",
"dense",
"(",
"x",
",",
"hidden_size",
",",
"name",
"=",
"\"tanh_discrete_unbottleneck\"",
")",
"return",
"x"
] | Simple un-discretization from tanh. | [
"Simple",
"un",
"-",
"discretization",
"from",
"tanh",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1393-L1396 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | isemhash_bottleneck | def isemhash_bottleneck(x,
bottleneck_bits,
bottleneck_noise,
discretize_warmup_steps,
mode,
isemhash_noise_dev=0.5,
isemhash_mix_prob=0.5):
"""Improved semantic hashing bott... | python | def isemhash_bottleneck(x,
bottleneck_bits,
bottleneck_noise,
discretize_warmup_steps,
mode,
isemhash_noise_dev=0.5,
isemhash_mix_prob=0.5):
"""Improved semantic hashing bott... | [
"def",
"isemhash_bottleneck",
"(",
"x",
",",
"bottleneck_bits",
",",
"bottleneck_noise",
",",
"discretize_warmup_steps",
",",
"mode",
",",
"isemhash_noise_dev",
"=",
"0.5",
",",
"isemhash_mix_prob",
"=",
"0.5",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
... | Improved semantic hashing bottleneck. | [
"Improved",
"semantic",
"hashing",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1399-L1426 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | isemhash_unbottleneck | def isemhash_unbottleneck(x, hidden_size, isemhash_filter_size_multiplier=1.0):
"""Improved semantic hashing un-bottleneck."""
filter_size = int(hidden_size * isemhash_filter_size_multiplier)
x = 0.5 * (x - 1.0) # Move from [-1, 1] to [0, 1].
with tf.variable_scope("isemhash_unbottleneck"):
h1a = tf.layers... | python | def isemhash_unbottleneck(x, hidden_size, isemhash_filter_size_multiplier=1.0):
"""Improved semantic hashing un-bottleneck."""
filter_size = int(hidden_size * isemhash_filter_size_multiplier)
x = 0.5 * (x - 1.0) # Move from [-1, 1] to [0, 1].
with tf.variable_scope("isemhash_unbottleneck"):
h1a = tf.layers... | [
"def",
"isemhash_unbottleneck",
"(",
"x",
",",
"hidden_size",
",",
"isemhash_filter_size_multiplier",
"=",
"1.0",
")",
":",
"filter_size",
"=",
"int",
"(",
"hidden_size",
"*",
"isemhash_filter_size_multiplier",
")",
"x",
"=",
"0.5",
"*",
"(",
"x",
"-",
"1.0",
... | Improved semantic hashing un-bottleneck. | [
"Improved",
"semantic",
"hashing",
"un",
"-",
"bottleneck",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1429-L1437 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | parametrized_bottleneck | def parametrized_bottleneck(x, hparams):
"""Meta-function calling all the above bottlenecks with hparams."""
if hparams.bottleneck_kind == "tanh_discrete":
d, _ = tanh_discrete_bottleneck(
x, hparams.bottleneck_bits, hparams.bottleneck_noise * 0.5,
hparams.discretize_warmup_steps, hparams.mode)
... | python | def parametrized_bottleneck(x, hparams):
"""Meta-function calling all the above bottlenecks with hparams."""
if hparams.bottleneck_kind == "tanh_discrete":
d, _ = tanh_discrete_bottleneck(
x, hparams.bottleneck_bits, hparams.bottleneck_noise * 0.5,
hparams.discretize_warmup_steps, hparams.mode)
... | [
"def",
"parametrized_bottleneck",
"(",
"x",
",",
"hparams",
")",
":",
"if",
"hparams",
".",
"bottleneck_kind",
"==",
"\"tanh_discrete\"",
":",
"d",
",",
"_",
"=",
"tanh_discrete_bottleneck",
"(",
"x",
",",
"hparams",
".",
"bottleneck_bits",
",",
"hparams",
"."... | Meta-function calling all the above bottlenecks with hparams. | [
"Meta",
"-",
"function",
"calling",
"all",
"the",
"above",
"bottlenecks",
"with",
"hparams",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1440-L1476 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | parametrized_unbottleneck | def parametrized_unbottleneck(x, hidden_size, hparams):
"""Meta-function calling all the above un-bottlenecks with hparams."""
if hparams.bottleneck_kind == "tanh_discrete":
return tanh_discrete_unbottleneck(x, hidden_size)
if hparams.bottleneck_kind == "isemhash":
return isemhash_unbottleneck(x, hidden_s... | python | def parametrized_unbottleneck(x, hidden_size, hparams):
"""Meta-function calling all the above un-bottlenecks with hparams."""
if hparams.bottleneck_kind == "tanh_discrete":
return tanh_discrete_unbottleneck(x, hidden_size)
if hparams.bottleneck_kind == "isemhash":
return isemhash_unbottleneck(x, hidden_s... | [
"def",
"parametrized_unbottleneck",
"(",
"x",
",",
"hidden_size",
",",
"hparams",
")",
":",
"if",
"hparams",
".",
"bottleneck_kind",
"==",
"\"tanh_discrete\"",
":",
"return",
"tanh_discrete_unbottleneck",
"(",
"x",
",",
"hidden_size",
")",
"if",
"hparams",
".",
... | Meta-function calling all the above un-bottlenecks with hparams. | [
"Meta",
"-",
"function",
"calling",
"all",
"the",
"above",
"un",
"-",
"bottlenecks",
"with",
"hparams",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1479-L1489 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/discretization.py | iaf_hparams | def iaf_hparams(hidden_size=512, filter_size=4096):
"""Create hyperpameters for inverse autoregressive flows.
Args:
hidden_size: Width of attention layers and neural network output layer.
filter_size: Hidden layer width for neural network.
Returns:
hparams: Hyperpameters with basic presets for inver... | python | def iaf_hparams(hidden_size=512, filter_size=4096):
"""Create hyperpameters for inverse autoregressive flows.
Args:
hidden_size: Width of attention layers and neural network output layer.
filter_size: Hidden layer width for neural network.
Returns:
hparams: Hyperpameters with basic presets for inver... | [
"def",
"iaf_hparams",
"(",
"hidden_size",
"=",
"512",
",",
"filter_size",
"=",
"4096",
")",
":",
"hparams",
"=",
"common_hparams",
".",
"basic_params1",
"(",
")",
"# Attention hyperparameters.",
"hparams",
".",
"hidden_size",
"=",
"hidden_size",
"hparams",
".",
... | Create hyperpameters for inverse autoregressive flows.
Args:
hidden_size: Width of attention layers and neural network output layer.
filter_size: Hidden layer width for neural network.
Returns:
hparams: Hyperpameters with basic presets for inverse autoregressive flows. | [
"Create",
"hyperpameters",
"for",
"inverse",
"autoregressive",
"flows",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/discretization.py#L1492-L1528 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/lm1b.py | _original_vocab | def _original_vocab(tmp_dir):
"""Returns a set containing the original vocabulary.
This is important for comparing with published results.
Args:
tmp_dir: directory containing dataset.
Returns:
a set of strings
"""
vocab_url = ("http://download.tensorflow.org/models/LM_LSTM_CNN/"
"v... | python | def _original_vocab(tmp_dir):
"""Returns a set containing the original vocabulary.
This is important for comparing with published results.
Args:
tmp_dir: directory containing dataset.
Returns:
a set of strings
"""
vocab_url = ("http://download.tensorflow.org/models/LM_LSTM_CNN/"
"v... | [
"def",
"_original_vocab",
"(",
"tmp_dir",
")",
":",
"vocab_url",
"=",
"(",
"\"http://download.tensorflow.org/models/LM_LSTM_CNN/\"",
"\"vocab-2016-09-10.txt\"",
")",
"vocab_filename",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"vocab_url",
"+",
"\".en\"",
")",
"voc... | Returns a set containing the original vocabulary.
This is important for comparing with published results.
Args:
tmp_dir: directory containing dataset.
Returns:
a set of strings | [
"Returns",
"a",
"set",
"containing",
"the",
"original",
"vocabulary",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/lm1b.py#L35-L55 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/lm1b.py | _replace_oov | def _replace_oov(original_vocab, line):
"""Replace out-of-vocab words with "UNK".
This maintains compatibility with published results.
Args:
original_vocab: a set of strings (The standard vocabulary for the dataset)
line: a unicode string - a space-delimited sequence of words.
Returns:
a unicode ... | python | def _replace_oov(original_vocab, line):
"""Replace out-of-vocab words with "UNK".
This maintains compatibility with published results.
Args:
original_vocab: a set of strings (The standard vocabulary for the dataset)
line: a unicode string - a space-delimited sequence of words.
Returns:
a unicode ... | [
"def",
"_replace_oov",
"(",
"original_vocab",
",",
"line",
")",
":",
"return",
"u\" \"",
".",
"join",
"(",
"[",
"word",
"if",
"word",
"in",
"original_vocab",
"else",
"u\"UNK\"",
"for",
"word",
"in",
"line",
".",
"split",
"(",
")",
"]",
")"
] | Replace out-of-vocab words with "UNK".
This maintains compatibility with published results.
Args:
original_vocab: a set of strings (The standard vocabulary for the dataset)
line: a unicode string - a space-delimited sequence of words.
Returns:
a unicode string - a space-delimited sequence of words. | [
"Replace",
"out",
"-",
"of",
"-",
"vocab",
"words",
"with",
"UNK",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/lm1b.py#L58-L71 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/lm1b.py | _maybe_download_corpus | def _maybe_download_corpus(tmp_dir):
"""Download and unpack the corpus.
Args:
tmp_dir: directory containing dataset.
"""
corpus_url = ("http://www.statmt.org/lm-benchmark/"
"1-billion-word-language-modeling-benchmark-r13output.tar.gz")
corpus_filename = os.path.basename(corpus_url)
corp... | python | def _maybe_download_corpus(tmp_dir):
"""Download and unpack the corpus.
Args:
tmp_dir: directory containing dataset.
"""
corpus_url = ("http://www.statmt.org/lm-benchmark/"
"1-billion-word-language-modeling-benchmark-r13output.tar.gz")
corpus_filename = os.path.basename(corpus_url)
corp... | [
"def",
"_maybe_download_corpus",
"(",
"tmp_dir",
")",
":",
"corpus_url",
"=",
"(",
"\"http://www.statmt.org/lm-benchmark/\"",
"\"1-billion-word-language-modeling-benchmark-r13output.tar.gz\"",
")",
"corpus_filename",
"=",
"os",
".",
"path",
".",
"basename",
"(",
"corpus_url",... | Download and unpack the corpus.
Args:
tmp_dir: directory containing dataset. | [
"Download",
"and",
"unpack",
"the",
"corpus",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/lm1b.py#L90-L103 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/cycle_gan.py | lossfn | def lossfn(real_input, fake_input, compress, hparams, lsgan, name):
"""Loss function."""
eps = 1e-12
with tf.variable_scope(name):
d1 = discriminator(real_input, compress, hparams, "discriminator")
d2 = discriminator(fake_input, compress, hparams, "discriminator",
reuse=True)
if... | python | def lossfn(real_input, fake_input, compress, hparams, lsgan, name):
"""Loss function."""
eps = 1e-12
with tf.variable_scope(name):
d1 = discriminator(real_input, compress, hparams, "discriminator")
d2 = discriminator(fake_input, compress, hparams, "discriminator",
reuse=True)
if... | [
"def",
"lossfn",
"(",
"real_input",
",",
"fake_input",
",",
"compress",
",",
"hparams",
",",
"lsgan",
",",
"name",
")",
":",
"eps",
"=",
"1e-12",
"with",
"tf",
".",
"variable_scope",
"(",
"name",
")",
":",
"d1",
"=",
"discriminator",
"(",
"real_input",
... | Loss function. | [
"Loss",
"function",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/cycle_gan.py#L46-L63 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/cycle_gan.py | cycle_gan_internal | def cycle_gan_internal(inputs, targets, _, hparams):
"""Cycle GAN, main step used for training."""
with tf.variable_scope("cycle_gan"):
# Embed inputs and targets.
inputs_orig, targets_orig = tf.to_int32(inputs), tf.to_int32(targets)
inputs = common_layers.embedding(
inputs_orig, hparams.vocab_s... | python | def cycle_gan_internal(inputs, targets, _, hparams):
"""Cycle GAN, main step used for training."""
with tf.variable_scope("cycle_gan"):
# Embed inputs and targets.
inputs_orig, targets_orig = tf.to_int32(inputs), tf.to_int32(targets)
inputs = common_layers.embedding(
inputs_orig, hparams.vocab_s... | [
"def",
"cycle_gan_internal",
"(",
"inputs",
",",
"targets",
",",
"_",
",",
"hparams",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"\"cycle_gan\"",
")",
":",
"# Embed inputs and targets.",
"inputs_orig",
",",
"targets_orig",
"=",
"tf",
".",
"to_int32",
... | Cycle GAN, main step used for training. | [
"Cycle",
"GAN",
"main",
"step",
"used",
"for",
"training",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/cycle_gan.py#L72-L113 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/cycle_gan.py | cycle_gan_small | def cycle_gan_small():
"""Set of hyperparameters."""
hparams = transformer_vae.transformer_ae_small()
hparams.batch_size = 2048
hparams.bottom = {
"inputs": modalities.identity_bottom,
"targets": modalities.identity_bottom,
}
hparams.top = {
"targets": modalities.identity_top,
}
hparam... | python | def cycle_gan_small():
"""Set of hyperparameters."""
hparams = transformer_vae.transformer_ae_small()
hparams.batch_size = 2048
hparams.bottom = {
"inputs": modalities.identity_bottom,
"targets": modalities.identity_bottom,
}
hparams.top = {
"targets": modalities.identity_top,
}
hparam... | [
"def",
"cycle_gan_small",
"(",
")",
":",
"hparams",
"=",
"transformer_vae",
".",
"transformer_ae_small",
"(",
")",
"hparams",
".",
"batch_size",
"=",
"2048",
"hparams",
".",
"bottom",
"=",
"{",
"\"inputs\"",
":",
"modalities",
".",
"identity_bottom",
",",
"\"t... | Set of hyperparameters. | [
"Set",
"of",
"hyperparameters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/cycle_gan.py#L126-L144 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | decode_hparams | def decode_hparams(overrides=""):
"""Hparams for decoding."""
hparams = decoding.decode_hparams()
# Number of interpolations between [0.0, 1.0].
hparams.add_hparam("num_interp", 11)
# Which level(s) to interpolate.
hparams.add_hparam("level_interp", [0, 1, 2])
# "all" or "ranked", interpolate all channels... | python | def decode_hparams(overrides=""):
"""Hparams for decoding."""
hparams = decoding.decode_hparams()
# Number of interpolations between [0.0, 1.0].
hparams.add_hparam("num_interp", 11)
# Which level(s) to interpolate.
hparams.add_hparam("level_interp", [0, 1, 2])
# "all" or "ranked", interpolate all channels... | [
"def",
"decode_hparams",
"(",
"overrides",
"=",
"\"\"",
")",
":",
"hparams",
"=",
"decoding",
".",
"decode_hparams",
"(",
")",
"# Number of interpolations between [0.0, 1.0].",
"hparams",
".",
"add_hparam",
"(",
"\"num_interp\"",
",",
"11",
")",
"# Which level(s) to i... | Hparams for decoding. | [
"Hparams",
"for",
"decoding",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L53-L67 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | preprocess_frame | def preprocess_frame(frame):
"""Preprocess frame.
1. Converts [0, 255] to [-0.5, 0.5]
2. Adds uniform noise.
Args:
frame: 3-D Tensor representing pixels.
Returns:
frame: 3-D Tensor with values in between [-0.5, 0.5]
"""
# Normalize from [0.0, 1.0] -> [-0.5, 0.5]
frame = common_layers.convert_r... | python | def preprocess_frame(frame):
"""Preprocess frame.
1. Converts [0, 255] to [-0.5, 0.5]
2. Adds uniform noise.
Args:
frame: 3-D Tensor representing pixels.
Returns:
frame: 3-D Tensor with values in between [-0.5, 0.5]
"""
# Normalize from [0.0, 1.0] -> [-0.5, 0.5]
frame = common_layers.convert_r... | [
"def",
"preprocess_frame",
"(",
"frame",
")",
":",
"# Normalize from [0.0, 1.0] -> [-0.5, 0.5]",
"frame",
"=",
"common_layers",
".",
"convert_rgb_to_real",
"(",
"frame",
")",
"frame",
"=",
"frame",
"-",
"0.5",
"frame",
",",
"_",
"=",
"glow_ops",
".",
"uniform_binn... | Preprocess frame.
1. Converts [0, 255] to [-0.5, 0.5]
2. Adds uniform noise.
Args:
frame: 3-D Tensor representing pixels.
Returns:
frame: 3-D Tensor with values in between [-0.5, 0.5] | [
"Preprocess",
"frame",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L70-L85 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | frame_to_latents | def frame_to_latents(frame, hparams):
"""Encode frames to latents."""
# Preprocess
frame = preprocess_frame(frame)
# Encode [X_t] to [z^1_t, z^2_t .. z^l_t]
glow_vals = glow_ops.encoder_decoder(
"codec", frame, hparams, eps=None, reverse=False)
z_top, _, level_eps, _, _ = glow_vals
return z_top, le... | python | def frame_to_latents(frame, hparams):
"""Encode frames to latents."""
# Preprocess
frame = preprocess_frame(frame)
# Encode [X_t] to [z^1_t, z^2_t .. z^l_t]
glow_vals = glow_ops.encoder_decoder(
"codec", frame, hparams, eps=None, reverse=False)
z_top, _, level_eps, _, _ = glow_vals
return z_top, le... | [
"def",
"frame_to_latents",
"(",
"frame",
",",
"hparams",
")",
":",
"# Preprocess",
"frame",
"=",
"preprocess_frame",
"(",
"frame",
")",
"# Encode [X_t] to [z^1_t, z^2_t .. z^l_t]",
"glow_vals",
"=",
"glow_ops",
".",
"encoder_decoder",
"(",
"\"codec\"",
",",
"frame",
... | Encode frames to latents. | [
"Encode",
"frames",
"to",
"latents",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L88-L97 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | latents_to_frames | def latents_to_frames(z_top_interp, level_eps_interp, hparams):
"""Decodes latents to frames."""
# Decode [z^1_t, z^2_t .. z^l_t] to [X_t]
images, _, _, _ = glow_ops.encoder_decoder(
"codec", z_top_interp, hparams, eps=level_eps_interp, reverse=True)
images = glow_ops.postprocess(images)
return images | python | def latents_to_frames(z_top_interp, level_eps_interp, hparams):
"""Decodes latents to frames."""
# Decode [z^1_t, z^2_t .. z^l_t] to [X_t]
images, _, _, _ = glow_ops.encoder_decoder(
"codec", z_top_interp, hparams, eps=level_eps_interp, reverse=True)
images = glow_ops.postprocess(images)
return images | [
"def",
"latents_to_frames",
"(",
"z_top_interp",
",",
"level_eps_interp",
",",
"hparams",
")",
":",
"# Decode [z^1_t, z^2_t .. z^l_t] to [X_t]",
"images",
",",
"_",
",",
"_",
",",
"_",
"=",
"glow_ops",
".",
"encoder_decoder",
"(",
"\"codec\"",
",",
"z_top_interp",
... | Decodes latents to frames. | [
"Decodes",
"latents",
"to",
"frames",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L100-L106 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | interpolate | def interpolate(features, hparams, decode_hp):
"""Interpolate between the first input frame and last target frame.
Args:
features: dict of tensors
hparams: HParams, training hparams.
decode_hp: HParams, decode hparams.
Returns:
images: interpolated images, 4-D Tensor, shape=(num_interp, H, W, C)
... | python | def interpolate(features, hparams, decode_hp):
"""Interpolate between the first input frame and last target frame.
Args:
features: dict of tensors
hparams: HParams, training hparams.
decode_hp: HParams, decode hparams.
Returns:
images: interpolated images, 4-D Tensor, shape=(num_interp, H, W, C)
... | [
"def",
"interpolate",
"(",
"features",
",",
"hparams",
",",
"decode_hp",
")",
":",
"inputs",
",",
"targets",
"=",
"features",
"[",
"\"inputs\"",
"]",
",",
"features",
"[",
"\"targets\"",
"]",
"inputs",
"=",
"tf",
".",
"unstack",
"(",
"inputs",
",",
"axis... | Interpolate between the first input frame and last target frame.
Args:
features: dict of tensors
hparams: HParams, training hparams.
decode_hp: HParams, decode hparams.
Returns:
images: interpolated images, 4-D Tensor, shape=(num_interp, H, W, C)
first_frame: image, 3-D Tensor, shape=(1, H, W, ... | [
"Interpolate",
"between",
"the",
"first",
"input",
"frame",
"and",
"last",
"target",
"frame",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L109-L150 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | get_summaries_log_dir | def get_summaries_log_dir(decode_hp, output_dir, dataset_split):
"""Get nested summaries_log_dir based on decode_hp."""
child_dir = decode_hp.summaries_log_dir
level_dir = "".join([str(level) for level in decode_hp.level_interp])
if decode_hp.channel_interp == "all":
rank_dir = "all"
else:
rank_dir = ... | python | def get_summaries_log_dir(decode_hp, output_dir, dataset_split):
"""Get nested summaries_log_dir based on decode_hp."""
child_dir = decode_hp.summaries_log_dir
level_dir = "".join([str(level) for level in decode_hp.level_interp])
if decode_hp.channel_interp == "all":
rank_dir = "all"
else:
rank_dir = ... | [
"def",
"get_summaries_log_dir",
"(",
"decode_hp",
",",
"output_dir",
",",
"dataset_split",
")",
":",
"child_dir",
"=",
"decode_hp",
".",
"summaries_log_dir",
"level_dir",
"=",
"\"\"",
".",
"join",
"(",
"[",
"str",
"(",
"level",
")",
"for",
"level",
"in",
"de... | Get nested summaries_log_dir based on decode_hp. | [
"Get",
"nested",
"summaries_log_dir",
"based",
"on",
"decode_hp",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L153-L164 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/nfg_interpolate.py | interpolations_to_summary | def interpolations_to_summary(sample_ind, interpolations, first_frame,
last_frame, hparams, decode_hp):
"""Converts interpolated frames into tf summaries.
The summaries consists of:
1. Image summary corresponding to the first frame.
2. Image summary corresponding to the last f... | python | def interpolations_to_summary(sample_ind, interpolations, first_frame,
last_frame, hparams, decode_hp):
"""Converts interpolated frames into tf summaries.
The summaries consists of:
1. Image summary corresponding to the first frame.
2. Image summary corresponding to the last f... | [
"def",
"interpolations_to_summary",
"(",
"sample_ind",
",",
"interpolations",
",",
"first_frame",
",",
"last_frame",
",",
"hparams",
",",
"decode_hp",
")",
":",
"parent_tag",
"=",
"\"sample_%d\"",
"%",
"sample_ind",
"frame_shape",
"=",
"hparams",
".",
"problem",
"... | Converts interpolated frames into tf summaries.
The summaries consists of:
1. Image summary corresponding to the first frame.
2. Image summary corresponding to the last frame.
3. The interpolated frames as a gif summary.
Args:
sample_ind: int
interpolations: Numpy array, shape=(num_interp, H, ... | [
"Converts",
"interpolated",
"frames",
"into",
"tf",
"summaries",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/nfg_interpolate.py#L167-L205 | train |
tensorflow/tensor2tensor | tensor2tensor/models/video/epva_params.py | next_frame_epva | def next_frame_epva():
"""EPVA hparams."""
hparams = basic_deterministic_params.next_frame_basic_deterministic()
hparams.video_num_input_frames = 4
hparams.video_num_target_frames = 4
hparams.bottom = {
"inputs": modalities.video_raw_bottom,
"targets": modalities.video_raw_targets_bottom,
}
hp... | python | def next_frame_epva():
"""EPVA hparams."""
hparams = basic_deterministic_params.next_frame_basic_deterministic()
hparams.video_num_input_frames = 4
hparams.video_num_target_frames = 4
hparams.bottom = {
"inputs": modalities.video_raw_bottom,
"targets": modalities.video_raw_targets_bottom,
}
hp... | [
"def",
"next_frame_epva",
"(",
")",
":",
"hparams",
"=",
"basic_deterministic_params",
".",
"next_frame_basic_deterministic",
"(",
")",
"hparams",
".",
"video_num_input_frames",
"=",
"4",
"hparams",
".",
"video_num_target_frames",
"=",
"4",
"hparams",
".",
"bottom",
... | EPVA hparams. | [
"EPVA",
"hparams",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/video/epva_params.py#L27-L63 | train |
tensorflow/tensor2tensor | tensor2tensor/utils/multistep_optimizer.py | MultistepAdamOptimizer._create_slots | def _create_slots(self, var_list):
"""Create slot variables for Adam with accumulated gradients."""
super(MultistepAdamOptimizer, self)._create_slots(var_list)
first_var = min(var_list, key=lambda x: x.name)
self._create_non_slot_variable(initial_value=0 if self._n == 1 else 1,
... | python | def _create_slots(self, var_list):
"""Create slot variables for Adam with accumulated gradients."""
super(MultistepAdamOptimizer, self)._create_slots(var_list)
first_var = min(var_list, key=lambda x: x.name)
self._create_non_slot_variable(initial_value=0 if self._n == 1 else 1,
... | [
"def",
"_create_slots",
"(",
"self",
",",
"var_list",
")",
":",
"super",
"(",
"MultistepAdamOptimizer",
",",
"self",
")",
".",
"_create_slots",
"(",
"var_list",
")",
"first_var",
"=",
"min",
"(",
"var_list",
",",
"key",
"=",
"lambda",
"x",
":",
"x",
".",... | Create slot variables for Adam with accumulated gradients. | [
"Create",
"slot",
"variables",
"for",
"Adam",
"with",
"accumulated",
"gradients",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/utils/multistep_optimizer.py#L43-L51 | train |
tensorflow/tensor2tensor | tensor2tensor/utils/multistep_optimizer.py | MultistepAdamOptimizer._apply_cond | def _apply_cond(self, apply_fn, grad, var, *args, **kwargs):
"""Apply conditionally if counter is zero."""
grad_acc = self.get_slot(var, "grad_acc")
def apply_adam(grad_acc, apply_fn, grad, var, *args, **kwargs):
total_grad = (grad_acc + grad) / tf.cast(self._n_t, grad.dtype)
adam_op = apply_fn... | python | def _apply_cond(self, apply_fn, grad, var, *args, **kwargs):
"""Apply conditionally if counter is zero."""
grad_acc = self.get_slot(var, "grad_acc")
def apply_adam(grad_acc, apply_fn, grad, var, *args, **kwargs):
total_grad = (grad_acc + grad) / tf.cast(self._n_t, grad.dtype)
adam_op = apply_fn... | [
"def",
"_apply_cond",
"(",
"self",
",",
"apply_fn",
",",
"grad",
",",
"var",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"grad_acc",
"=",
"self",
".",
"get_slot",
"(",
"var",
",",
"\"grad_acc\"",
")",
"def",
"apply_adam",
"(",
"grad_acc",
",... | Apply conditionally if counter is zero. | [
"Apply",
"conditionally",
"if",
"counter",
"is",
"zero",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/utils/multistep_optimizer.py#L62-L81 | train |
tensorflow/tensor2tensor | tensor2tensor/utils/multistep_optimizer.py | MultistepAdamOptimizer._finish | def _finish(self, update_ops, name_scope):
"""Updates beta_power variables every n batches and incrs counter."""
iter_ = self._get_iter_variable()
beta1_power, beta2_power = self._get_beta_accumulators()
with tf.control_dependencies(update_ops):
with tf.colocate_with(iter_):
def update_be... | python | def _finish(self, update_ops, name_scope):
"""Updates beta_power variables every n batches and incrs counter."""
iter_ = self._get_iter_variable()
beta1_power, beta2_power = self._get_beta_accumulators()
with tf.control_dependencies(update_ops):
with tf.colocate_with(iter_):
def update_be... | [
"def",
"_finish",
"(",
"self",
",",
"update_ops",
",",
"name_scope",
")",
":",
"iter_",
"=",
"self",
".",
"_get_iter_variable",
"(",
")",
"beta1_power",
",",
"beta2_power",
"=",
"self",
".",
"_get_beta_accumulators",
"(",
")",
"with",
"tf",
".",
"control_dep... | Updates beta_power variables every n batches and incrs counter. | [
"Updates",
"beta_power",
"variables",
"every",
"n",
"batches",
"and",
"incrs",
"counter",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/utils/multistep_optimizer.py#L103-L124 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/transformer_revnet.py | transformer_revnet_encoder | def transformer_revnet_encoder(encoder_input,
encoder_self_attention_bias,
hparams,
name="encoder"):
"""A stack of transformer layers.
Args:
encoder_input: a Tensor
encoder_self_attention_bias: bias Tensor for self... | python | def transformer_revnet_encoder(encoder_input,
encoder_self_attention_bias,
hparams,
name="encoder"):
"""A stack of transformer layers.
Args:
encoder_input: a Tensor
encoder_self_attention_bias: bias Tensor for self... | [
"def",
"transformer_revnet_encoder",
"(",
"encoder_input",
",",
"encoder_self_attention_bias",
",",
"hparams",
",",
"name",
"=",
"\"encoder\"",
")",
":",
"def",
"f",
"(",
"x",
",",
"side_input",
")",
":",
"\"\"\"f(x) for reversible layer, self-attention layer.\"\"\"",
"... | A stack of transformer layers.
Args:
encoder_input: a Tensor
encoder_self_attention_bias: bias Tensor for self-attention
(see common_attention.attention_bias())
hparams: hyperparameters for model
name: a string
Returns:
y: a Tensors | [
"A",
"stack",
"of",
"transformer",
"layers",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/transformer_revnet.py#L73-L133 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/transformer_revnet.py | transformer_revnet_decoder | def transformer_revnet_decoder(decoder_input,
encoder_output,
decoder_self_attention_bias,
encoder_decoder_attention_bias,
hparams,
name="decoder"):
"""A stack of ... | python | def transformer_revnet_decoder(decoder_input,
encoder_output,
decoder_self_attention_bias,
encoder_decoder_attention_bias,
hparams,
name="decoder"):
"""A stack of ... | [
"def",
"transformer_revnet_decoder",
"(",
"decoder_input",
",",
"encoder_output",
",",
"decoder_self_attention_bias",
",",
"encoder_decoder_attention_bias",
",",
"hparams",
",",
"name",
"=",
"\"decoder\"",
")",
":",
"def",
"f",
"(",
"x",
",",
"side_input",
")",
":",... | A stack of transformer layers.
Args:
decoder_input: a Tensor
encoder_output: a Tensor
decoder_self_attention_bias: bias Tensor for self-attention
(see common_attention.attention_bias())
encoder_decoder_attention_bias: bias Tensor for encoder-decoder attention
(see common_attention.attenti... | [
"A",
"stack",
"of",
"transformer",
"layers",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/transformer_revnet.py#L136-L213 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/transformer_revnet.py | transformer_revnet_base | def transformer_revnet_base():
"""Base hparams for TransformerRevnet."""
hparams = transformer.transformer_big()
# Use settings from transformer_n_da
hparams.layer_preprocess_sequence = "n"
hparams.layer_postprocess_sequence = "da"
hparams.learning_rate = 0.4
return hparams | python | def transformer_revnet_base():
"""Base hparams for TransformerRevnet."""
hparams = transformer.transformer_big()
# Use settings from transformer_n_da
hparams.layer_preprocess_sequence = "n"
hparams.layer_postprocess_sequence = "da"
hparams.learning_rate = 0.4
return hparams | [
"def",
"transformer_revnet_base",
"(",
")",
":",
"hparams",
"=",
"transformer",
".",
"transformer_big",
"(",
")",
"# Use settings from transformer_n_da",
"hparams",
".",
"layer_preprocess_sequence",
"=",
"\"n\"",
"hparams",
".",
"layer_postprocess_sequence",
"=",
"\"da\""... | Base hparams for TransformerRevnet. | [
"Base",
"hparams",
"for",
"TransformerRevnet",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/transformer_revnet.py#L217-L226 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/transformer_revnet.py | transformer_revnet_big | def transformer_revnet_big():
"""Base hparams for TransformerRevnet."""
hparams = transformer_revnet_base()
# The TransformerRevnet uses significantly less memory than the Transformer.
# Increase batch size and model size.
hparams.batch_size *= 2
hparams.hidden_size *= 2
hparams.num_heads *= 2
hparams.... | python | def transformer_revnet_big():
"""Base hparams for TransformerRevnet."""
hparams = transformer_revnet_base()
# The TransformerRevnet uses significantly less memory than the Transformer.
# Increase batch size and model size.
hparams.batch_size *= 2
hparams.hidden_size *= 2
hparams.num_heads *= 2
hparams.... | [
"def",
"transformer_revnet_big",
"(",
")",
":",
"hparams",
"=",
"transformer_revnet_base",
"(",
")",
"# The TransformerRevnet uses significantly less memory than the Transformer.",
"# Increase batch size and model size.",
"hparams",
".",
"batch_size",
"*=",
"2",
"hparams",
".",
... | Base hparams for TransformerRevnet. | [
"Base",
"hparams",
"for",
"TransformerRevnet",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/transformer_revnet.py#L230-L240 | train |
tensorflow/tensor2tensor | tensor2tensor/utils/devices.py | data_parallelism_from_flags | def data_parallelism_from_flags(daisy_chain_variables=True, all_workers=False):
"""Over which devices do we split each training batch.
In old-fashioned async mode, we split the batch over all GPUs on the
current worker.
In sync mode, we split the batch over all the parameter server GPUs.
This function retu... | python | def data_parallelism_from_flags(daisy_chain_variables=True, all_workers=False):
"""Over which devices do we split each training batch.
In old-fashioned async mode, we split the batch over all GPUs on the
current worker.
In sync mode, we split the batch over all the parameter server GPUs.
This function retu... | [
"def",
"data_parallelism_from_flags",
"(",
"daisy_chain_variables",
"=",
"True",
",",
"all_workers",
"=",
"False",
")",
":",
"dp_arg_names",
"=",
"inspect",
".",
"getargspec",
"(",
"data_parallelism",
")",
".",
"args",
"blacklist",
"=",
"[",
"\"daisy_chain_variables... | Over which devices do we split each training batch.
In old-fashioned async mode, we split the batch over all GPUs on the
current worker.
In sync mode, we split the batch over all the parameter server GPUs.
This function returns an expert_utils.Parallelism object, which can be used
to build the model. It i... | [
"Over",
"which",
"devices",
"do",
"we",
"split",
"each",
"training",
"batch",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/utils/devices.py#L26-L59 | train |
tensorflow/tensor2tensor | tensor2tensor/utils/devices.py | data_parallelism | def data_parallelism(daisy_chain_variables=True,
all_workers=False,
ps_replicas=0,
ps_job="/job:ps",
ps_gpu=0,
schedule="continuous_train_and_eval",
sync=False,
worker_gpu=1... | python | def data_parallelism(daisy_chain_variables=True,
all_workers=False,
ps_replicas=0,
ps_job="/job:ps",
ps_gpu=0,
schedule="continuous_train_and_eval",
sync=False,
worker_gpu=1... | [
"def",
"data_parallelism",
"(",
"daisy_chain_variables",
"=",
"True",
",",
"all_workers",
"=",
"False",
",",
"ps_replicas",
"=",
"0",
",",
"ps_job",
"=",
"\"/job:ps\"",
",",
"ps_gpu",
"=",
"0",
",",
"schedule",
"=",
"\"continuous_train_and_eval\"",
",",
"sync",
... | See data_parallelism_from_flags. | [
"See",
"data_parallelism_from_flags",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/utils/devices.py#L62-L177 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/wiki_lm.py | concat_generator | def concat_generator(filename, up_threshold, low_threshold=10):
"""Generate concatenated lines from file upto up_threshold characters."""
txt = ""
for line in tf.gfile.Open(filename):
line = line.strip()
if len(txt) + len(line) + 1 >= up_threshold:
ret = txt
txt = ""
# We don't yield ver... | python | def concat_generator(filename, up_threshold, low_threshold=10):
"""Generate concatenated lines from file upto up_threshold characters."""
txt = ""
for line in tf.gfile.Open(filename):
line = line.strip()
if len(txt) + len(line) + 1 >= up_threshold:
ret = txt
txt = ""
# We don't yield ver... | [
"def",
"concat_generator",
"(",
"filename",
",",
"up_threshold",
",",
"low_threshold",
"=",
"10",
")",
":",
"txt",
"=",
"\"\"",
"for",
"line",
"in",
"tf",
".",
"gfile",
".",
"Open",
"(",
"filename",
")",
":",
"line",
"=",
"line",
".",
"strip",
"(",
"... | Generate concatenated lines from file upto up_threshold characters. | [
"Generate",
"concatenated",
"lines",
"from",
"file",
"upto",
"up_threshold",
"characters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/wiki_lm.py#L33-L48 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/wiki_lm.py | mix_generators | def mix_generators(generator_list):
"""Given python generators, generate from one, then from another, etc."""
i = 0
l = len(generator_list)
stopiters_seen = 0
while stopiters_seen <= l:
try:
yield six.next(generator_list[i % l])
i += 1
stopiters_seen = 0
except StopIteration:
i... | python | def mix_generators(generator_list):
"""Given python generators, generate from one, then from another, etc."""
i = 0
l = len(generator_list)
stopiters_seen = 0
while stopiters_seen <= l:
try:
yield six.next(generator_list[i % l])
i += 1
stopiters_seen = 0
except StopIteration:
i... | [
"def",
"mix_generators",
"(",
"generator_list",
")",
":",
"i",
"=",
"0",
"l",
"=",
"len",
"(",
"generator_list",
")",
"stopiters_seen",
"=",
"0",
"while",
"stopiters_seen",
"<=",
"l",
":",
"try",
":",
"yield",
"six",
".",
"next",
"(",
"generator_list",
"... | Given python generators, generate from one, then from another, etc. | [
"Given",
"python",
"generators",
"generate",
"from",
"one",
"then",
"from",
"another",
"etc",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/wiki_lm.py#L51-L63 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/translate.py | compute_bleu_summaries | def compute_bleu_summaries(hook_args):
"""Compute BLEU core summaries using the decoder output.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
A list of tf.Summary values if hook_args.hparams contains the
reference file and the translated file.
"""
decode_hparams = hook_args.decode_hparams
... | python | def compute_bleu_summaries(hook_args):
"""Compute BLEU core summaries using the decoder output.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
A list of tf.Summary values if hook_args.hparams contains the
reference file and the translated file.
"""
decode_hparams = hook_args.decode_hparams
... | [
"def",
"compute_bleu_summaries",
"(",
"hook_args",
")",
":",
"decode_hparams",
"=",
"hook_args",
".",
"decode_hparams",
"if",
"not",
"(",
"decode_hparams",
".",
"decode_reference",
"and",
"decode_hparams",
".",
"decode_to_file",
")",
":",
"return",
"None",
"values",... | Compute BLEU core summaries using the decoder output.
Args:
hook_args: DecodeHookArgs namedtuple
Returns:
A list of tf.Summary values if hook_args.hparams contains the
reference file and the translated file. | [
"Compute",
"BLEU",
"core",
"summaries",
"using",
"the",
"decoder",
"output",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/translate.py#L83-L118 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/translate.py | _preprocess_sgm | def _preprocess_sgm(line, is_sgm):
"""Preprocessing to strip tags in SGM files."""
if not is_sgm:
return line
# In SGM files, remove <srcset ...>, <p>, <doc ...> lines.
if line.startswith("<srcset") or line.startswith("</srcset"):
return ""
if line.startswith("<doc") or line.startswith("</doc"):
r... | python | def _preprocess_sgm(line, is_sgm):
"""Preprocessing to strip tags in SGM files."""
if not is_sgm:
return line
# In SGM files, remove <srcset ...>, <p>, <doc ...> lines.
if line.startswith("<srcset") or line.startswith("</srcset"):
return ""
if line.startswith("<doc") or line.startswith("</doc"):
r... | [
"def",
"_preprocess_sgm",
"(",
"line",
",",
"is_sgm",
")",
":",
"if",
"not",
"is_sgm",
":",
"return",
"line",
"# In SGM files, remove <srcset ...>, <p>, <doc ...> lines.",
"if",
"line",
".",
"startswith",
"(",
"\"<srcset\"",
")",
"or",
"line",
".",
"startswith",
"... | Preprocessing to strip tags in SGM files. | [
"Preprocessing",
"to",
"strip",
"tags",
"in",
"SGM",
"files",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/translate.py#L121-L136 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/translate.py | compile_data | def compile_data(tmp_dir, datasets, filename, datatypes_to_clean=None):
"""Concatenates all `datasets` and saves to `filename`."""
datatypes_to_clean = datatypes_to_clean or []
filename = os.path.join(tmp_dir, filename)
lang1_fname = filename + ".lang1"
lang2_fname = filename + ".lang2"
if tf.gfile.Exists(l... | python | def compile_data(tmp_dir, datasets, filename, datatypes_to_clean=None):
"""Concatenates all `datasets` and saves to `filename`."""
datatypes_to_clean = datatypes_to_clean or []
filename = os.path.join(tmp_dir, filename)
lang1_fname = filename + ".lang1"
lang2_fname = filename + ".lang2"
if tf.gfile.Exists(l... | [
"def",
"compile_data",
"(",
"tmp_dir",
",",
"datasets",
",",
"filename",
",",
"datatypes_to_clean",
"=",
"None",
")",
":",
"datatypes_to_clean",
"=",
"datatypes_to_clean",
"or",
"[",
"]",
"filename",
"=",
"os",
".",
"path",
".",
"join",
"(",
"tmp_dir",
",",
... | Concatenates all `datasets` and saves to `filename`. | [
"Concatenates",
"all",
"datasets",
"and",
"saves",
"to",
"filename",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/translate.py#L158-L255 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/translate.py | TranslateDistillProblem.get_or_create_vocab | def get_or_create_vocab(self, data_dir, tmp_dir, force_get=False):
"""Get vocab for distill problems."""
# We assume that vocab file is present in data_dir directory where the
# data generated will be stored.
vocab_filepath = os.path.join(data_dir, self.vocab_filename)
encoder = text_encoder.Subword... | python | def get_or_create_vocab(self, data_dir, tmp_dir, force_get=False):
"""Get vocab for distill problems."""
# We assume that vocab file is present in data_dir directory where the
# data generated will be stored.
vocab_filepath = os.path.join(data_dir, self.vocab_filename)
encoder = text_encoder.Subword... | [
"def",
"get_or_create_vocab",
"(",
"self",
",",
"data_dir",
",",
"tmp_dir",
",",
"force_get",
"=",
"False",
")",
":",
"# We assume that vocab file is present in data_dir directory where the",
"# data generated will be stored.",
"vocab_filepath",
"=",
"os",
".",
"path",
".",... | Get vocab for distill problems. | [
"Get",
"vocab",
"for",
"distill",
"problems",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/translate.py#L278-L284 | train |
tensorflow/tensor2tensor | tensor2tensor/bin/t2t_trainer.py | set_hparams_from_args | def set_hparams_from_args(args):
"""Set hparams overrides from unparsed args list."""
if not args:
return
hp_prefix = "--hp_"
tf.logging.info("Found unparsed command-line arguments. Checking if any "
"start with %s and interpreting those as hparams "
"settings.", hp_pref... | python | def set_hparams_from_args(args):
"""Set hparams overrides from unparsed args list."""
if not args:
return
hp_prefix = "--hp_"
tf.logging.info("Found unparsed command-line arguments. Checking if any "
"start with %s and interpreting those as hparams "
"settings.", hp_pref... | [
"def",
"set_hparams_from_args",
"(",
"args",
")",
":",
"if",
"not",
"args",
":",
"return",
"hp_prefix",
"=",
"\"--hp_\"",
"tf",
".",
"logging",
".",
"info",
"(",
"\"Found unparsed command-line arguments. Checking if any \"",
"\"start with %s and interpreting those as hparam... | Set hparams overrides from unparsed args list. | [
"Set",
"hparams",
"overrides",
"from",
"unparsed",
"args",
"list",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/bin/t2t_trainer.py#L138-L162 | train |
tensorflow/tensor2tensor | tensor2tensor/bin/t2t_trainer.py | create_hparams | def create_hparams():
"""Create hparams."""
if FLAGS.use_tpu and "tpu" not in FLAGS.hparams_set:
tf.logging.warn("Not all hyperparameter sets work on TPU. "
"Prefer hparams_sets with a '_tpu' suffix, "
"e.g. transformer_tpu, if available for your model.")
hparams_path =... | python | def create_hparams():
"""Create hparams."""
if FLAGS.use_tpu and "tpu" not in FLAGS.hparams_set:
tf.logging.warn("Not all hyperparameter sets work on TPU. "
"Prefer hparams_sets with a '_tpu' suffix, "
"e.g. transformer_tpu, if available for your model.")
hparams_path =... | [
"def",
"create_hparams",
"(",
")",
":",
"if",
"FLAGS",
".",
"use_tpu",
"and",
"\"tpu\"",
"not",
"in",
"FLAGS",
".",
"hparams_set",
":",
"tf",
".",
"logging",
".",
"warn",
"(",
"\"Not all hyperparameter sets work on TPU. \"",
"\"Prefer hparams_sets with a '_tpu' suffix... | Create hparams. | [
"Create",
"hparams",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/bin/t2t_trainer.py#L165-L173 | train |
tensorflow/tensor2tensor | tensor2tensor/bin/t2t_trainer.py | create_run_config | def create_run_config(hp, output_dir=None):
"""Create a run config.
Args:
hp: model hyperparameters
output_dir: model's output directory, defaults to output_dir flag.
Returns:
a run config
"""
save_ckpt_steps = max(FLAGS.iterations_per_loop, FLAGS.local_eval_frequency)
save_ckpt_secs = FLAGS.s... | python | def create_run_config(hp, output_dir=None):
"""Create a run config.
Args:
hp: model hyperparameters
output_dir: model's output directory, defaults to output_dir flag.
Returns:
a run config
"""
save_ckpt_steps = max(FLAGS.iterations_per_loop, FLAGS.local_eval_frequency)
save_ckpt_secs = FLAGS.s... | [
"def",
"create_run_config",
"(",
"hp",
",",
"output_dir",
"=",
"None",
")",
":",
"save_ckpt_steps",
"=",
"max",
"(",
"FLAGS",
".",
"iterations_per_loop",
",",
"FLAGS",
".",
"local_eval_frequency",
")",
"save_ckpt_secs",
"=",
"FLAGS",
".",
"save_checkpoints_secs",
... | Create a run config.
Args:
hp: model hyperparameters
output_dir: model's output directory, defaults to output_dir flag.
Returns:
a run config | [
"Create",
"a",
"run",
"config",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/bin/t2t_trainer.py#L207-L275 | train |
tensorflow/tensor2tensor | tensor2tensor/bin/t2t_trainer.py | save_metadata | def save_metadata(hparams):
"""Saves FLAGS and hparams to output_dir."""
output_dir = os.path.expanduser(FLAGS.output_dir)
if not tf.gfile.Exists(output_dir):
tf.gfile.MakeDirs(output_dir)
# Save FLAGS in txt file
if hasattr(FLAGS, "flags_into_string"):
flags_str = FLAGS.flags_into_string()
t2t_f... | python | def save_metadata(hparams):
"""Saves FLAGS and hparams to output_dir."""
output_dir = os.path.expanduser(FLAGS.output_dir)
if not tf.gfile.Exists(output_dir):
tf.gfile.MakeDirs(output_dir)
# Save FLAGS in txt file
if hasattr(FLAGS, "flags_into_string"):
flags_str = FLAGS.flags_into_string()
t2t_f... | [
"def",
"save_metadata",
"(",
"hparams",
")",
":",
"output_dir",
"=",
"os",
".",
"path",
".",
"expanduser",
"(",
"FLAGS",
".",
"output_dir",
")",
"if",
"not",
"tf",
".",
"gfile",
".",
"Exists",
"(",
"output_dir",
")",
":",
"tf",
".",
"gfile",
".",
"Ma... | Saves FLAGS and hparams to output_dir. | [
"Saves",
"FLAGS",
"and",
"hparams",
"to",
"output_dir",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/bin/t2t_trainer.py#L313-L348 | train |
tensorflow/tensor2tensor | tensor2tensor/models/xception.py | residual_block | def residual_block(x, hparams):
"""A stack of convolution blocks with residual connection."""
k = (hparams.kernel_height, hparams.kernel_width)
dilations_and_kernels = [((1, 1), k) for _ in range(3)]
y = common_layers.subseparable_conv_block(
x,
hparams.hidden_size,
dilations_and_kernels,
... | python | def residual_block(x, hparams):
"""A stack of convolution blocks with residual connection."""
k = (hparams.kernel_height, hparams.kernel_width)
dilations_and_kernels = [((1, 1), k) for _ in range(3)]
y = common_layers.subseparable_conv_block(
x,
hparams.hidden_size,
dilations_and_kernels,
... | [
"def",
"residual_block",
"(",
"x",
",",
"hparams",
")",
":",
"k",
"=",
"(",
"hparams",
".",
"kernel_height",
",",
"hparams",
".",
"kernel_width",
")",
"dilations_and_kernels",
"=",
"[",
"(",
"(",
"1",
",",
"1",
")",
",",
"k",
")",
"for",
"_",
"in",
... | A stack of convolution blocks with residual connection. | [
"A",
"stack",
"of",
"convolution",
"blocks",
"with",
"residual",
"connection",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/xception.py#L33-L45 | train |
tensorflow/tensor2tensor | tensor2tensor/models/xception.py | xception_internal | def xception_internal(inputs, hparams):
"""Xception body."""
with tf.variable_scope("xception"):
cur = inputs
if cur.get_shape().as_list()[1] > 200:
# Large image, Xception entry flow
cur = xception_entry(cur, hparams.hidden_size)
else:
# Small image, conv
cur = common_layers.co... | python | def xception_internal(inputs, hparams):
"""Xception body."""
with tf.variable_scope("xception"):
cur = inputs
if cur.get_shape().as_list()[1] > 200:
# Large image, Xception entry flow
cur = xception_entry(cur, hparams.hidden_size)
else:
# Small image, conv
cur = common_layers.co... | [
"def",
"xception_internal",
"(",
"inputs",
",",
"hparams",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"\"xception\"",
")",
":",
"cur",
"=",
"inputs",
"if",
"cur",
".",
"get_shape",
"(",
")",
".",
"as_list",
"(",
")",
"[",
"1",
"]",
">",
"200... | Xception body. | [
"Xception",
"body",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/xception.py#L48-L70 | train |
tensorflow/tensor2tensor | tensor2tensor/models/xception.py | xception_entry | def xception_entry(inputs, hidden_dim):
"""Xception entry flow."""
with tf.variable_scope("xception_entry"):
def xnet_resblock(x, filters, res_relu, name):
"""Resblock."""
with tf.variable_scope(name):
y = common_layers.separable_conv_block(
x,
filters, [((1, 1), (3,... | python | def xception_entry(inputs, hidden_dim):
"""Xception entry flow."""
with tf.variable_scope("xception_entry"):
def xnet_resblock(x, filters, res_relu, name):
"""Resblock."""
with tf.variable_scope(name):
y = common_layers.separable_conv_block(
x,
filters, [((1, 1), (3,... | [
"def",
"xception_entry",
"(",
"inputs",
",",
"hidden_dim",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"\"xception_entry\"",
")",
":",
"def",
"xnet_resblock",
"(",
"x",
",",
"filters",
",",
"res_relu",
",",
"name",
")",
":",
"\"\"\"Resblock.\"\"\"",
... | Xception entry flow. | [
"Xception",
"entry",
"flow",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/xception.py#L73-L110 | train |
tensorflow/tensor2tensor | tensor2tensor/models/xception.py | xception_exit | def xception_exit(inputs):
"""Xception exit flow."""
with tf.variable_scope("xception_exit"):
x = inputs
x_shape = x.get_shape().as_list()
if x_shape[1] is None or x_shape[2] is None:
length_float = tf.to_float(tf.shape(x)[1])
length_float *= tf.to_float(tf.shape(x)[2])
spatial_dim_flo... | python | def xception_exit(inputs):
"""Xception exit flow."""
with tf.variable_scope("xception_exit"):
x = inputs
x_shape = x.get_shape().as_list()
if x_shape[1] is None or x_shape[2] is None:
length_float = tf.to_float(tf.shape(x)[1])
length_float *= tf.to_float(tf.shape(x)[2])
spatial_dim_flo... | [
"def",
"xception_exit",
"(",
"inputs",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"\"xception_exit\"",
")",
":",
"x",
"=",
"inputs",
"x_shape",
"=",
"x",
".",
"get_shape",
"(",
")",
".",
"as_list",
"(",
")",
"if",
"x_shape",
"[",
"1",
"]",
"... | Xception exit flow. | [
"Xception",
"exit",
"flow",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/xception.py#L113-L133 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/wikisum/html.py | get_text_from_html | def get_text_from_html(html):
"""Returns a plaintext representation of HTML content."""
try:
soup = bs4.BeautifulSoup(html, "html.parser")
except: # pylint: disable=bare-except
# Some docs don't parse
return ""
# Remove script and style tags
for s in soup(["script", "style"]):
s.decompose()
... | python | def get_text_from_html(html):
"""Returns a plaintext representation of HTML content."""
try:
soup = bs4.BeautifulSoup(html, "html.parser")
except: # pylint: disable=bare-except
# Some docs don't parse
return ""
# Remove script and style tags
for s in soup(["script", "style"]):
s.decompose()
... | [
"def",
"get_text_from_html",
"(",
"html",
")",
":",
"try",
":",
"soup",
"=",
"bs4",
".",
"BeautifulSoup",
"(",
"html",
",",
"\"html.parser\"",
")",
"except",
":",
"# pylint: disable=bare-except",
"# Some docs don't parse",
"return",
"\"\"",
"# Remove script and style ... | Returns a plaintext representation of HTML content. | [
"Returns",
"a",
"plaintext",
"representation",
"of",
"HTML",
"content",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/wikisum/html.py#L21-L32 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/wikisum/html.py | _soup_strings | def _soup_strings(soup):
"""Return text strings in soup."""
paragraph_tags = set([
"caption", "details", "h1", "h2", "h3", "h4", "h5", "h6", "li", "p", "td",
"div", "span"
])
skip_children = None
for descendant in soup.descendants:
# If we've treated a tag as a contiguous paragraph, don't re-... | python | def _soup_strings(soup):
"""Return text strings in soup."""
paragraph_tags = set([
"caption", "details", "h1", "h2", "h3", "h4", "h5", "h6", "li", "p", "td",
"div", "span"
])
skip_children = None
for descendant in soup.descendants:
# If we've treated a tag as a contiguous paragraph, don't re-... | [
"def",
"_soup_strings",
"(",
"soup",
")",
":",
"paragraph_tags",
"=",
"set",
"(",
"[",
"\"caption\"",
",",
"\"details\"",
",",
"\"h1\"",
",",
"\"h2\"",
",",
"\"h3\"",
",",
"\"h4\"",
",",
"\"h5\"",
",",
"\"h6\"",
",",
"\"li\"",
",",
"\"p\"",
",",
"\"td\""... | Return text strings in soup. | [
"Return",
"text",
"strings",
"in",
"soup",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/wikisum/html.py#L35-L78 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | image_transformer_base | def image_transformer_base():
"""Set of hyperparameters."""
hparams = common_hparams.basic_params1()
hparams.hidden_size = 512
hparams.batch_size = 4
hparams.max_length = 3075
hparams.dropout = 0.0
hparams.clip_grad_norm = 0. # i.e. no gradient clipping
hparams.optimizer_adam_epsilon = 1e-9
hparams.l... | python | def image_transformer_base():
"""Set of hyperparameters."""
hparams = common_hparams.basic_params1()
hparams.hidden_size = 512
hparams.batch_size = 4
hparams.max_length = 3075
hparams.dropout = 0.0
hparams.clip_grad_norm = 0. # i.e. no gradient clipping
hparams.optimizer_adam_epsilon = 1e-9
hparams.l... | [
"def",
"image_transformer_base",
"(",
")",
":",
"hparams",
"=",
"common_hparams",
".",
"basic_params1",
"(",
")",
"hparams",
".",
"hidden_size",
"=",
"512",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"max_length",
"=",
"3075",
"hparams",
".",
"d... | Set of hyperparameters. | [
"Set",
"of",
"hyperparameters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L172-L245 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_cifar10_base | def imagetransformer_cifar10_base():
"""Best config for 2.90 bits/dim on CIFAR10 using cross entropy."""
hparams = image_transformer_base()
hparams.batch_size = 4
hparams.num_heads = 4
hparams.num_decoder_layers = 12
hparams.block_length = 256
hparams.hidden_size = 512
hparams.filter_size = 2048
hpara... | python | def imagetransformer_cifar10_base():
"""Best config for 2.90 bits/dim on CIFAR10 using cross entropy."""
hparams = image_transformer_base()
hparams.batch_size = 4
hparams.num_heads = 4
hparams.num_decoder_layers = 12
hparams.block_length = 256
hparams.hidden_size = 512
hparams.filter_size = 2048
hpara... | [
"def",
"imagetransformer_cifar10_base",
"(",
")",
":",
"hparams",
"=",
"image_transformer_base",
"(",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"4",
"hparams",
".",
"num_decoder_layers",
"=",
"12",
"hparams",
".",
"block_leng... | Best config for 2.90 bits/dim on CIFAR10 using cross entropy. | [
"Best",
"config",
"for",
"2",
".",
"90",
"bits",
"/",
"dim",
"on",
"CIFAR10",
"using",
"cross",
"entropy",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L255-L270 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_cifar10_base_dmol | def imagetransformer_cifar10_base_dmol():
"""Best config for 2.90 bits/dim on CIFAR10 using DMOL."""
hparams = image_transformer_base()
hparams.likelihood = cia.DistributionType.DMOL
hparams.num_channels = 1
hparams.bottom["targets"] = modalities.image_channel_compress_targets_bottom
hparams.top["targets"] ... | python | def imagetransformer_cifar10_base_dmol():
"""Best config for 2.90 bits/dim on CIFAR10 using DMOL."""
hparams = image_transformer_base()
hparams.likelihood = cia.DistributionType.DMOL
hparams.num_channels = 1
hparams.bottom["targets"] = modalities.image_channel_compress_targets_bottom
hparams.top["targets"] ... | [
"def",
"imagetransformer_cifar10_base_dmol",
"(",
")",
":",
"hparams",
"=",
"image_transformer_base",
"(",
")",
"hparams",
".",
"likelihood",
"=",
"cia",
".",
"DistributionType",
".",
"DMOL",
"hparams",
".",
"num_channels",
"=",
"1",
"hparams",
".",
"bottom",
"[... | Best config for 2.90 bits/dim on CIFAR10 using DMOL. | [
"Best",
"config",
"for",
"2",
".",
"90",
"bits",
"/",
"dim",
"on",
"CIFAR10",
"using",
"DMOL",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L274-L298 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_tpu | def imagetransformer_base_tpu():
"""Transformer base params for cifar-10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length = 128
... | python | def imagetransformer_base_tpu():
"""Transformer base params for cifar-10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length = 128
... | [
"def",
"imagetransformer_base_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"4",
"# heads are ... | Transformer base params for cifar-10. | [
"Transformer",
"base",
"params",
"for",
"cifar",
"-",
"10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L302-L317 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_imagenet_tpu | def imagetransformer_base_imagenet_tpu():
"""Transformer base params for cifar-10."""
hparams = imagetransformer_base_tpu()
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length = 128
hparams.layer_preprocess_sequence = "none"
hp... | python | def imagetransformer_base_imagenet_tpu():
"""Transformer base params for cifar-10."""
hparams = imagetransformer_base_tpu()
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length = 128
hparams.layer_preprocess_sequence = "none"
hp... | [
"def",
"imagetransformer_base_imagenet_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_tpu",
"(",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"4",
"# heads are expensive on tpu",
"hparams",
".",
"num_decoder_layers",
"=",... | Transformer base params for cifar-10. | [
"Transformer",
"base",
"params",
"for",
"cifar",
"-",
"10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L321-L331 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels | def imagetransformer_sep_channels():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 4
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = 512
hparams.num_hidden_layers = 6
return hparams | python | def imagetransformer_sep_channels():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 4
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = 512
hparams.num_hidden_layers = 6
return hparams | [
"def",
"imagetransformer_sep_channels",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base",
"(",
")",
"hparams",
".",
"num_heads",
"=",
"4",
"hparams",
".",
"attention_key_channels",
"=",
"hparams",
".",
"attention_value_channels",
"=",
"0",
"hparams",
".",
"h... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L352-L360 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_8l | def imagetransformer_sep_channels_8l():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 4
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = 256
hparams.num_hidden_layers = 8
hparams.sampling_method =... | python | def imagetransformer_sep_channels_8l():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 4
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = 256
hparams.num_hidden_layers = 8
hparams.sampling_method =... | [
"def",
"imagetransformer_sep_channels_8l",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base",
"(",
")",
"hparams",
".",
"num_heads",
"=",
"4",
"hparams",
".",
"attention_key_channels",
"=",
"hparams",
".",
"attention_value_channels",
"=",
"0",
"hparams",
".",
... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L364-L373 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_8l_8h_big_cond_dr03_dan | def imagetransformer_base_8l_8h_big_cond_dr03_dan():
"""big 1d model for conditional image generation.2.99 on cifar10."""
hparams = imagetransformer_sep_channels_8l()
hparams.block_width = 256
hparams.block_length = 256
hparams.hidden_size = 512
hparams.num_heads = 8
hparams.filter_size = 2048
hparams.b... | python | def imagetransformer_base_8l_8h_big_cond_dr03_dan():
"""big 1d model for conditional image generation.2.99 on cifar10."""
hparams = imagetransformer_sep_channels_8l()
hparams.block_width = 256
hparams.block_length = 256
hparams.hidden_size = 512
hparams.num_heads = 8
hparams.filter_size = 2048
hparams.b... | [
"def",
"imagetransformer_base_8l_8h_big_cond_dr03_dan",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_8l",
"(",
")",
"hparams",
".",
"block_width",
"=",
"256",
"hparams",
".",
"block_length",
"=",
"256",
"hparams",
".",
"hidden_size",
"=",
"512",
"hp... | big 1d model for conditional image generation.2.99 on cifar10. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
".",
"2",
".",
"99",
"on",
"cifar10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L386-L400 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_10l_8h_big_uncond_dr03_dan_64 | def imagetransformer_base_10l_8h_big_uncond_dr03_dan_64():
"""big 1d model for unconditional generation on imagenet."""
hparams = imagetransformer_base_10l_8h_big_cond_dr03_dan()
hparams.unconditional = True
hparams.max_length = 14000
hparams.batch_size = 1
hparams.img_len = 64
hparams.layer_prepostproces... | python | def imagetransformer_base_10l_8h_big_uncond_dr03_dan_64():
"""big 1d model for unconditional generation on imagenet."""
hparams = imagetransformer_base_10l_8h_big_cond_dr03_dan()
hparams.unconditional = True
hparams.max_length = 14000
hparams.batch_size = 1
hparams.img_len = 64
hparams.layer_prepostproces... | [
"def",
"imagetransformer_base_10l_8h_big_uncond_dr03_dan_64",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_10l_8h_big_cond_dr03_dan",
"(",
")",
"hparams",
".",
"unconditional",
"=",
"True",
"hparams",
".",
"max_length",
"=",
"14000",
"hparams",
".",
"batch_size",... | big 1d model for unconditional generation on imagenet. | [
"big",
"1d",
"model",
"for",
"unconditional",
"generation",
"on",
"imagenet",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L404-L412 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformerpp_sep_channels_8l_8h | def imagetransformerpp_sep_channels_8l_8h():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.likelihood = cia.DistributionType.DMOL
hparams.num_channels = 1
hparams.bottom["targets"] = modalities.image_channel_compress_targets_bottom
hparams.top["targets"] = modalities.identity_top
... | python | def imagetransformerpp_sep_channels_8l_8h():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.likelihood = cia.DistributionType.DMOL
hparams.num_channels = 1
hparams.bottom["targets"] = modalities.image_channel_compress_targets_bottom
hparams.top["targets"] = modalities.identity_top
... | [
"def",
"imagetransformerpp_sep_channels_8l_8h",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base",
"(",
")",
"hparams",
".",
"likelihood",
"=",
"cia",
".",
"DistributionType",
".",
"DMOL",
"hparams",
".",
"num_channels",
"=",
"1",
"hparams",
".",
"bottom",
... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L416-L434 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformerpp_base_8l_8h_big_cond_dr03_dan | def imagetransformerpp_base_8l_8h_big_cond_dr03_dan():
"""big 1d model for conditional image generation.2.99 on cifar10."""
hparams = imagetransformerpp_sep_channels_8l_8h()
hparams.hidden_size = 512
hparams.num_heads = 8
hparams.filter_size = 2048
hparams.batch_size = 4
hparams.max_length = 3075
hparam... | python | def imagetransformerpp_base_8l_8h_big_cond_dr03_dan():
"""big 1d model for conditional image generation.2.99 on cifar10."""
hparams = imagetransformerpp_sep_channels_8l_8h()
hparams.hidden_size = 512
hparams.num_heads = 8
hparams.filter_size = 2048
hparams.batch_size = 4
hparams.max_length = 3075
hparam... | [
"def",
"imagetransformerpp_base_8l_8h_big_cond_dr03_dan",
"(",
")",
":",
"hparams",
"=",
"imagetransformerpp_sep_channels_8l_8h",
"(",
")",
"hparams",
".",
"hidden_size",
"=",
"512",
"hparams",
".",
"num_heads",
"=",
"8",
"hparams",
".",
"filter_size",
"=",
"2048",
... | big 1d model for conditional image generation.2.99 on cifar10. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
".",
"2",
".",
"99",
"on",
"cifar10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L438-L451 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformerpp_base_14l_8h_big_uncond_dr03_dan_p | def imagetransformerpp_base_14l_8h_big_uncond_dr03_dan_p():
"""Gets to 2.92 in just under 4 days on 8 p100s."""
hparams = imagetransformerpp_base_12l_8h_big_uncond_dr03_dan_l()
hparams.num_decoder_layers = 14
hparams.batch_size = 8
hparams.layer_prepostprocess_dropout = 0.2
return hparams | python | def imagetransformerpp_base_14l_8h_big_uncond_dr03_dan_p():
"""Gets to 2.92 in just under 4 days on 8 p100s."""
hparams = imagetransformerpp_base_12l_8h_big_uncond_dr03_dan_l()
hparams.num_decoder_layers = 14
hparams.batch_size = 8
hparams.layer_prepostprocess_dropout = 0.2
return hparams | [
"def",
"imagetransformerpp_base_14l_8h_big_uncond_dr03_dan_p",
"(",
")",
":",
"hparams",
"=",
"imagetransformerpp_base_12l_8h_big_uncond_dr03_dan_l",
"(",
")",
"hparams",
".",
"num_decoder_layers",
"=",
"14",
"hparams",
".",
"batch_size",
"=",
"8",
"hparams",
".",
"layer_... | Gets to 2.92 in just under 4 days on 8 p100s. | [
"Gets",
"to",
"2",
".",
"92",
"in",
"just",
"under",
"4",
"days",
"on",
"8",
"p100s",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L537-L543 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformerpp_base_5l_8h_big_uncond_dr00_dan_g_bs1 | def imagetransformerpp_base_5l_8h_big_uncond_dr00_dan_g_bs1():
"""For 256x256."""
hparams = imagetransformerpp_base_10l_8h_big_uncond_dr03_dan_g()
# TODO(trandustin): I forgot to set this in the runs! Maybe it's not used in
# image transformer training implementation?
# hparams.img_len = 256
hparams.max_len... | python | def imagetransformerpp_base_5l_8h_big_uncond_dr00_dan_g_bs1():
"""For 256x256."""
hparams = imagetransformerpp_base_10l_8h_big_uncond_dr03_dan_g()
# TODO(trandustin): I forgot to set this in the runs! Maybe it's not used in
# image transformer training implementation?
# hparams.img_len = 256
hparams.max_len... | [
"def",
"imagetransformerpp_base_5l_8h_big_uncond_dr00_dan_g_bs1",
"(",
")",
":",
"hparams",
"=",
"imagetransformerpp_base_10l_8h_big_uncond_dr03_dan_g",
"(",
")",
"# TODO(trandustin): I forgot to set this in the runs! Maybe it's not used in",
"# image transformer training implementation?",
"... | For 256x256. | [
"For",
"256x256",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L565-L579 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_8l_8h_big_cond_dr03_dan_dilated | def imagetransformer_base_8l_8h_big_cond_dr03_dan_dilated():
"""Dilated hparams."""
hparams = imagetransformer_base_8l_8h_big_cond_dr03_dan()
hparams.gap_sizes = [0, 16, 64, 0, 16, 64, 128, 0]
hparams.dec_attention_type = cia.AttentionType.DILATED
hparams.block_length = 128
hparams.block_width = 128
hpara... | python | def imagetransformer_base_8l_8h_big_cond_dr03_dan_dilated():
"""Dilated hparams."""
hparams = imagetransformer_base_8l_8h_big_cond_dr03_dan()
hparams.gap_sizes = [0, 16, 64, 0, 16, 64, 128, 0]
hparams.dec_attention_type = cia.AttentionType.DILATED
hparams.block_length = 128
hparams.block_width = 128
hpara... | [
"def",
"imagetransformer_base_8l_8h_big_cond_dr03_dan_dilated",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_8l_8h_big_cond_dr03_dan",
"(",
")",
"hparams",
".",
"gap_sizes",
"=",
"[",
"0",
",",
"16",
",",
"64",
",",
"0",
",",
"16",
",",
"64",
",",
"128"... | Dilated hparams. | [
"Dilated",
"hparams",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L635-L643 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_12l_8h_big | def imagetransformer_base_12l_8h_big():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.filter_size = 1024
hparams.num_decoder_layers = 12
hparams.batch_size = 1
hparams.hidden_size = 512
hparams.learning_rate_warmup_steps = 4000
hparams.sampl... | python | def imagetransformer_base_12l_8h_big():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.filter_size = 1024
hparams.num_decoder_layers = 12
hparams.batch_size = 1
hparams.hidden_size = 512
hparams.learning_rate_warmup_steps = 4000
hparams.sampl... | [
"def",
"imagetransformer_base_12l_8h_big",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_8l_8h",
"(",
")",
"hparams",
".",
"filter_size",
"=",
"1024",
"hparams",
".",
"num_decoder_layers",
"=",
"12",
"hparams",
".",
"batch_size",
"=",
"1",
"hparams",... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L673-L684 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer1d_base_8l_64by64 | def imagetransformer1d_base_8l_64by64():
"""hparams fo 12 layer big 1d model for imagenet 64x64."""
hparams = image_transformer_base()
hparams.num_heads = 8
hparams.hidden_size = 512
hparams.filter_size = 2048
hparams.num_decoder_layers = 8
hparams.batch_size = 1
hparams.block_length = 512
hparams.blo... | python | def imagetransformer1d_base_8l_64by64():
"""hparams fo 12 layer big 1d model for imagenet 64x64."""
hparams = image_transformer_base()
hparams.num_heads = 8
hparams.hidden_size = 512
hparams.filter_size = 2048
hparams.num_decoder_layers = 8
hparams.batch_size = 1
hparams.block_length = 512
hparams.blo... | [
"def",
"imagetransformer1d_base_8l_64by64",
"(",
")",
":",
"hparams",
"=",
"image_transformer_base",
"(",
")",
"hparams",
".",
"num_heads",
"=",
"8",
"hparams",
".",
"hidden_size",
"=",
"512",
"hparams",
".",
"filter_size",
"=",
"2048",
"hparams",
".",
"num_deco... | hparams fo 12 layer big 1d model for imagenet 64x64. | [
"hparams",
"fo",
"12",
"layer",
"big",
"1d",
"model",
"for",
"imagenet",
"64x64",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L688-L701 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_12l_16h_imagenet_large | def imagetransformer_sep_channels_12l_16h_imagenet_large():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.num_hidden_layers = 12
hparams.batch_size = 1
hparams.filter_size = 2048
hparams.num_heads = 16
hparams.learning_rate_warmup_steps = 16000
hparams.sampling_m... | python | def imagetransformer_sep_channels_12l_16h_imagenet_large():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.num_hidden_layers = 12
hparams.batch_size = 1
hparams.filter_size = 2048
hparams.num_heads = 16
hparams.learning_rate_warmup_steps = 16000
hparams.sampling_m... | [
"def",
"imagetransformer_sep_channels_12l_16h_imagenet_large",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_8l_8h",
"(",
")",
"hparams",
".",
"num_hidden_layers",
"=",
"12",
"hparams",
".",
"batch_size",
"=",
"1",
"hparams",
".",
"filter_size",
"=",
"... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L754-L764 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc | def imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_12l_16h_imagenet_large()
hparams.num_hidden_layers = 16
hparams.local_attention = True
hparams.batch_size = 1
hparams.block_length = 256
return hparams | python | def imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_12l_16h_imagenet_large()
hparams.num_hidden_layers = 16
hparams.local_attention = True
hparams.batch_size = 1
hparams.block_length = 256
return hparams | [
"def",
"imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_12l_16h_imagenet_large",
"(",
")",
"hparams",
".",
"num_hidden_layers",
"=",
"16",
"hparams",
".",
"local_attention",
"=",
"True",
"hparams",
".",
... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L768-L775 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc_128 | def imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc_128():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_12l_16h_imagenet_large()
hparams.num_hidden_layers = 16
hparams.local_attention = True
hparams.batch_size = 1
hparams.block_length = 128
return hparams | python | def imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc_128():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_12l_16h_imagenet_large()
hparams.num_hidden_layers = 16
hparams.local_attention = True
hparams.batch_size = 1
hparams.block_length = 128
return hparams | [
"def",
"imagetransformer_sep_channels_16l_16h_imgnet_lrg_loc_128",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_12l_16h_imagenet_large",
"(",
")",
"hparams",
".",
"num_hidden_layers",
"=",
"16",
"hparams",
".",
"local_attention",
"=",
"True",
"hparams",
"."... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L779-L786 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_10l_16h_big_uncond_dr01_imgnet | def imagetransformer_base_10l_16h_big_uncond_dr01_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 16
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batc... | python | def imagetransformer_base_10l_16h_big_uncond_dr01_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 16
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batc... | [
"def",
"imagetransformer_base_10l_16h_big_uncond_dr01_imgnet",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_14l_8h_big_dr01",
"(",
")",
"# num_hidden_layers",
"hparams",
".",
"num_decoder_layers",
"=",
"10",
"hparams",
".",
"num_heads",
"=",
"16",
"hparams",
".",... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L799-L809 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_10l_16h_big_dr01_imgnet | def imagetransformer_base_10l_16h_big_dr01_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 16
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batch_size ... | python | def imagetransformer_base_10l_16h_big_dr01_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 16
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batch_size ... | [
"def",
"imagetransformer_base_10l_16h_big_dr01_imgnet",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_14l_8h_big_dr01",
"(",
")",
"# num_hidden_layers",
"hparams",
".",
"num_decoder_layers",
"=",
"10",
"hparams",
".",
"num_heads",
"=",
"16",
"hparams",
".",
"hid... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L813-L824 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_8l_8h | def imagetransformer_sep_channels_8l_8h():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 8
hparams.batch_size = 1
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 512
hparams.filter_size = 512
hparams.num_hidden_layers = 8... | python | def imagetransformer_sep_channels_8l_8h():
"""separate rgb embeddings."""
hparams = imagetransformer_base()
hparams.num_heads = 8
hparams.batch_size = 1
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 512
hparams.filter_size = 512
hparams.num_hidden_layers = 8... | [
"def",
"imagetransformer_sep_channels_8l_8h",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base",
"(",
")",
"hparams",
".",
"num_heads",
"=",
"8",
"hparams",
".",
"batch_size",
"=",
"1",
"hparams",
".",
"attention_key_channels",
"=",
"hparams",
".",
"attention... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L828-L838 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_8l_8h_local_and_global_att | def imagetransformer_sep_channels_8l_8h_local_and_global_att():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.num_heads = 8
hparams.batch_size = 1
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = ... | python | def imagetransformer_sep_channels_8l_8h_local_and_global_att():
"""separate rgb embeddings."""
hparams = imagetransformer_sep_channels_8l_8h()
hparams.num_heads = 8
hparams.batch_size = 1
hparams.attention_key_channels = hparams.attention_value_channels = 0
hparams.hidden_size = 256
hparams.filter_size = ... | [
"def",
"imagetransformer_sep_channels_8l_8h_local_and_global_att",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_8l_8h",
"(",
")",
"hparams",
".",
"num_heads",
"=",
"8",
"hparams",
".",
"batch_size",
"=",
"1",
"hparams",
".",
"attention_key_channels",
"=... | separate rgb embeddings. | [
"separate",
"rgb",
"embeddings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L842-L853 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_bas8l_8h_big_uncond_dr03_imgnet | def imagetransformer_bas8l_8h_big_uncond_dr03_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 8
hparams.num_heads = 8
hparams.hidden_size = 512
hparams.filter_size = 2048
hparams.layer_prepo... | python | def imagetransformer_bas8l_8h_big_uncond_dr03_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_14l_8h_big_dr01()
# num_hidden_layers
hparams.num_decoder_layers = 8
hparams.num_heads = 8
hparams.hidden_size = 512
hparams.filter_size = 2048
hparams.layer_prepo... | [
"def",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_14l_8h_big_dr01",
"(",
")",
"# num_hidden_layers",
"hparams",
".",
"num_decoder_layers",
"=",
"8",
"hparams",
".",
"num_heads",
"=",
"8",
"hparams",
".",
"hi... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L857-L866 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_base_10l_16h_big_dr01_moe_imgnet | def imagetransformer_base_10l_16h_big_dr01_moe_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_10l_16h_big_dr01_imgnet()
hparams.initializer = "orthogonal"
hparams.learning_rate_warmup_steps = 16000
hparams.add_hparam("moe_layers_decoder", "2,7") # Which layer i... | python | def imagetransformer_base_10l_16h_big_dr01_moe_imgnet():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_base_10l_16h_big_dr01_imgnet()
hparams.initializer = "orthogonal"
hparams.learning_rate_warmup_steps = 16000
hparams.add_hparam("moe_layers_decoder", "2,7") # Which layer i... | [
"def",
"imagetransformer_base_10l_16h_big_dr01_moe_imgnet",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_base_10l_16h_big_dr01_imgnet",
"(",
")",
"hparams",
".",
"initializer",
"=",
"\"orthogonal\"",
"hparams",
".",
"learning_rate_warmup_steps",
"=",
"16000",
"hparams",
... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L902-L914 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_moe_tiny | def imagetransformer_moe_tiny():
"""Set of hyperparameters for a very small imagetransformer with MoE."""
hparams = imagetransformer_tiny()
hparams.hidden_size = 64
hparams.batch_size = 1
hparams.num_hidden_layers = 3
hparams.dec_attention_type = cia.AttentionType.MOE_LOCAL_1D
hparams.add_hparam("moe_laye... | python | def imagetransformer_moe_tiny():
"""Set of hyperparameters for a very small imagetransformer with MoE."""
hparams = imagetransformer_tiny()
hparams.hidden_size = 64
hparams.batch_size = 1
hparams.num_hidden_layers = 3
hparams.dec_attention_type = cia.AttentionType.MOE_LOCAL_1D
hparams.add_hparam("moe_laye... | [
"def",
"imagetransformer_moe_tiny",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_tiny",
"(",
")",
"hparams",
".",
"hidden_size",
"=",
"64",
"hparams",
".",
"batch_size",
"=",
"1",
"hparams",
".",
"num_hidden_layers",
"=",
"3",
"hparams",
".",
"dec_attention_... | Set of hyperparameters for a very small imagetransformer with MoE. | [
"Set",
"of",
"hyperparameters",
"for",
"a",
"very",
"small",
"imagetransformer",
"with",
"MoE",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L918-L930 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_sep_channels_8l_tpu | def imagetransformer_sep_channels_8l_tpu():
"""Hparams for training imagetransformer on tpu."""
hparams = imagetransformer_sep_channels_8l()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.shared_embedding_and_softmax_weights = False
retu... | python | def imagetransformer_sep_channels_8l_tpu():
"""Hparams for training imagetransformer on tpu."""
hparams = imagetransformer_sep_channels_8l()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.shared_embedding_and_softmax_weights = False
retu... | [
"def",
"imagetransformer_sep_channels_8l_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_sep_channels_8l",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"4",
"# heads are expen... | Hparams for training imagetransformer on tpu. | [
"Hparams",
"for",
"training",
"imagetransformer",
"on",
"tpu",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L941-L948 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b10l_4h_big_uncond_dr03_tpu | def imagetransformer_b10l_4h_big_uncond_dr03_tpu():
"""Small model for tpu cifar 10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.block_len... | python | def imagetransformer_b10l_4h_big_uncond_dr03_tpu():
"""Small model for tpu cifar 10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.block_len... | [
"def",
"imagetransformer_b10l_4h_big_uncond_dr03_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"... | Small model for tpu cifar 10. | [
"Small",
"model",
"for",
"tpu",
"cifar",
"10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L952-L965 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b10l_dr03_moe_tpu | def imagetransformer_b10l_dr03_moe_tpu():
"""Moe tpu params."""
hparams = imagetransformer_b10l_4h_big_uncond_dr03_tpu()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.layer_preprocess_sequence = "none"
... | python | def imagetransformer_b10l_dr03_moe_tpu():
"""Moe tpu params."""
hparams = imagetransformer_b10l_4h_big_uncond_dr03_tpu()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.layer_preprocess_sequence = "none"
... | [
"def",
"imagetransformer_b10l_dr03_moe_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_b10l_4h_big_uncond_dr03_tpu",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
"4",
"# heads... | Moe tpu params. | [
"Moe",
"tpu",
"params",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L969-L979 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b10l_4h_big_uncond_dr03_lr025_tpu | def imagetransformer_b10l_4h_big_uncond_dr03_lr025_tpu():
"""TPU related small model."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.learning... | python | def imagetransformer_b10l_4h_big_uncond_dr03_lr025_tpu():
"""TPU related small model."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 10
hparams.learning... | [
"def",
"imagetransformer_b10l_4h_big_uncond_dr03_lr025_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=... | TPU related small model. | [
"TPU",
"related",
"small",
"model",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L983-L995 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_4h_b256_uncond_dr03_tpu | def imagetransformer_b12l_4h_b256_uncond_dr03_tpu():
"""works very well on 4x4."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length =... | python | def imagetransformer_b12l_4h_b256_uncond_dr03_tpu():
"""works very well on 4x4."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length =... | [
"def",
"imagetransformer_b12l_4h_b256_uncond_dr03_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"num_heads",
"=",
... | works very well on 4x4. | [
"works",
"very",
"well",
"on",
"4x4",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1025-L1041 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_4h_b256_uncond_dr03_rel_tpu | def imagetransformer_b12l_4h_b256_uncond_dr03_rel_tpu():
"""works very well on 4x4."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
hparams.shared_rel = True
hparams.dec_attention_type = cia.AttentionType.RELATIVE_LOCAL_1D
return hparams | python | def imagetransformer_b12l_4h_b256_uncond_dr03_rel_tpu():
"""works very well on 4x4."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
hparams.shared_rel = True
hparams.dec_attention_type = cia.AttentionType.RELATIVE_LOCAL_1D
return hparams | [
"def",
"imagetransformer_b12l_4h_b256_uncond_dr03_rel_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_b12l_4h_b256_uncond_dr03_tpu",
"(",
")",
"hparams",
".",
"shared_rel",
"=",
"True",
"hparams",
".",
"dec_attention_type",
"=",
"cia",
".",
"AttentionType",
".",
... | works very well on 4x4. | [
"works",
"very",
"well",
"on",
"4x4",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1045-L1050 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_cifar_tpu_range | def imagetransformer_cifar_tpu_range(rhp):
"""Range of hyperparameters for vizier."""
# After starting from base, set intervals for some parameters.
rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE)
rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16])
rhp.set_discrete("hidden_size", [256, ... | python | def imagetransformer_cifar_tpu_range(rhp):
"""Range of hyperparameters for vizier."""
# After starting from base, set intervals for some parameters.
rhp.set_float("learning_rate", 0.01, 1.0, scale=rhp.LOG_SCALE)
rhp.set_discrete("num_decoder_layers", [8, 10, 12, 14, 16])
rhp.set_discrete("hidden_size", [256, ... | [
"def",
"imagetransformer_cifar_tpu_range",
"(",
"rhp",
")",
":",
"# After starting from base, set intervals for some parameters.",
"rhp",
".",
"set_float",
"(",
"\"learning_rate\"",
",",
"0.01",
",",
"1.0",
",",
"scale",
"=",
"rhp",
".",
"LOG_SCALE",
")",
"rhp",
".",
... | Range of hyperparameters for vizier. | [
"Range",
"of",
"hyperparameters",
"for",
"vizier",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1054-L1062 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_4h_b128_h512_uncond_dr01_im | def imagetransformer_b12l_4h_b128_h512_uncond_dr01_im():
"""TPU related imagenet model."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.optimizer = "Adafactor"
hparams.learning_rate_schedule = "rsqrt_decay"
hparams.learning_rate_w... | python | def imagetransformer_b12l_4h_b128_h512_uncond_dr01_im():
"""TPU related imagenet model."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
update_hparams_for_tpu(hparams)
hparams.batch_size = 4
hparams.optimizer = "Adafactor"
hparams.learning_rate_schedule = "rsqrt_decay"
hparams.learning_rate_w... | [
"def",
"imagetransformer_b12l_4h_b128_h512_uncond_dr01_im",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_b12l_4h_b256_uncond_dr03_tpu",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"4",
"hparams",
".",
"optimizer",
"=",
... | TPU related imagenet model. | [
"TPU",
"related",
"imagenet",
"model",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1085-L1094 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_4h_uncond_dr03_tpu | def imagetransformer_b12l_4h_uncond_dr03_tpu():
"""TPU related small model."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
hparams.learning_rate = 0.2
hparams.learning_rate_warmup_steps = 4000
hparams.layer_preprocess_sequence = "none"
hparams.layer_postprocess_sequence = "dan"
hparams.layer... | python | def imagetransformer_b12l_4h_uncond_dr03_tpu():
"""TPU related small model."""
hparams = imagetransformer_b12l_4h_b256_uncond_dr03_tpu()
hparams.learning_rate = 0.2
hparams.learning_rate_warmup_steps = 4000
hparams.layer_preprocess_sequence = "none"
hparams.layer_postprocess_sequence = "dan"
hparams.layer... | [
"def",
"imagetransformer_b12l_4h_uncond_dr03_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_b12l_4h_b256_uncond_dr03_tpu",
"(",
")",
"hparams",
".",
"learning_rate",
"=",
"0.2",
"hparams",
".",
"learning_rate_warmup_steps",
"=",
"4000",
"hparams",
".",
"layer_prep... | TPU related small model. | [
"TPU",
"related",
"small",
"model",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1098-L1106 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_4h_b128_uncond_dr03_tpu | def imagetransformer_b12l_4h_b128_uncond_dr03_tpu():
"""TPU config for cifar 10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 2
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length ... | python | def imagetransformer_b12l_4h_b128_uncond_dr03_tpu():
"""TPU config for cifar 10."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 2
hparams.num_heads = 4 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.block_length ... | [
"def",
"imagetransformer_b12l_4h_b128_uncond_dr03_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"2",
"hparams",
".",
"num_heads",
"=",
... | TPU config for cifar 10. | [
"TPU",
"config",
"for",
"cifar",
"10",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1110-L1126 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b12l_8h_b256_uncond_dr03_tpu | def imagetransformer_b12l_8h_b256_uncond_dr03_tpu():
"""TPU related 12 layer 8 heads model."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 2
hparams.num_heads = 8 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.bl... | python | def imagetransformer_b12l_8h_b256_uncond_dr03_tpu():
"""TPU related 12 layer 8 heads model."""
hparams = imagetransformer_bas8l_8h_big_uncond_dr03_imgnet()
update_hparams_for_tpu(hparams)
hparams.batch_size = 2
hparams.num_heads = 8 # heads are expensive on tpu
hparams.num_decoder_layers = 12
hparams.bl... | [
"def",
"imagetransformer_b12l_8h_b256_uncond_dr03_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_bas8l_8h_big_uncond_dr03_imgnet",
"(",
")",
"update_hparams_for_tpu",
"(",
"hparams",
")",
"hparams",
".",
"batch_size",
"=",
"2",
"hparams",
".",
"num_heads",
"=",
... | TPU related 12 layer 8 heads model. | [
"TPU",
"related",
"12",
"layer",
"8",
"heads",
"model",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1130-L1143 | train |
tensorflow/tensor2tensor | tensor2tensor/models/image_transformer.py | imagetransformer_b10l_4h_big_uncond_dr01_tpu | def imagetransformer_b10l_4h_big_uncond_dr01_tpu():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_b12l_4h_big_uncond_dr03_tpu()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 4
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batch_... | python | def imagetransformer_b10l_4h_big_uncond_dr01_tpu():
"""big 1d model for conditional image generation."""
hparams = imagetransformer_b12l_4h_big_uncond_dr03_tpu()
# num_hidden_layers
hparams.num_decoder_layers = 10
hparams.num_heads = 4
hparams.hidden_size = 1024
hparams.filter_size = 4096
hparams.batch_... | [
"def",
"imagetransformer_b10l_4h_big_uncond_dr01_tpu",
"(",
")",
":",
"hparams",
"=",
"imagetransformer_b12l_4h_big_uncond_dr03_tpu",
"(",
")",
"# num_hidden_layers",
"hparams",
".",
"num_decoder_layers",
"=",
"10",
"hparams",
".",
"num_heads",
"=",
"4",
"hparams",
".",
... | big 1d model for conditional image generation. | [
"big",
"1d",
"model",
"for",
"conditional",
"image",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/image_transformer.py#L1147-L1157 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.