repo stringlengths 7 55 | path stringlengths 4 223 | func_name stringlengths 1 134 | original_string stringlengths 75 104k | language stringclasses 1
value | code stringlengths 75 104k | code_tokens listlengths 19 28.4k | docstring stringlengths 1 46.9k | docstring_tokens listlengths 1 1.97k | sha stringlengths 40 40 | url stringlengths 87 315 | partition stringclasses 1
value |
|---|---|---|---|---|---|---|---|---|---|---|---|
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | ResidualFeedForward | def ResidualFeedForward(feature_depth,
feedforward_depth,
dropout,
mode):
"""Residual feed-forward layer with normalization at start."""
return layers.Residual(
layers.LayerNorm(),
layers.Dense(feedforward_depth),
layers.Relu(... | python | def ResidualFeedForward(feature_depth,
feedforward_depth,
dropout,
mode):
"""Residual feed-forward layer with normalization at start."""
return layers.Residual(
layers.LayerNorm(),
layers.Dense(feedforward_depth),
layers.Relu(... | [
"def",
"ResidualFeedForward",
"(",
"feature_depth",
",",
"feedforward_depth",
",",
"dropout",
",",
"mode",
")",
":",
"return",
"layers",
".",
"Residual",
"(",
"layers",
".",
"LayerNorm",
"(",
")",
",",
"layers",
".",
"Dense",
"(",
"feedforward_depth",
")",
"... | Residual feed-forward layer with normalization at start. | [
"Residual",
"feed",
"-",
"forward",
"layer",
"with",
"normalization",
"at",
"start",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L24-L36 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | EncoderLayer | def EncoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
mode):
"""Transformer encoder layer.
The input to the encoder is a pair (embedded source, mask) where
the mask is created from the original source to prevent attending
to t... | python | def EncoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
mode):
"""Transformer encoder layer.
The input to the encoder is a pair (embedded source, mask) where
the mask is created from the original source to prevent attending
to t... | [
"def",
"EncoderLayer",
"(",
"feature_depth",
",",
"feedforward_depth",
",",
"num_heads",
",",
"dropout",
",",
"mode",
")",
":",
"# The encoder block expects (activation, mask) as input and returns",
"# the new activations only, we add the mask back to output next.",
"encoder_block",
... | Transformer encoder layer.
The input to the encoder is a pair (embedded source, mask) where
the mask is created from the original source to prevent attending
to the padding part of the input.
Args:
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_head... | [
"Transformer",
"encoder",
"layer",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L39-L76 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | TransformerEncoder | def TransformerEncoder(vocab_size,
num_classes=10,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
max_len=2048,
... | python | def TransformerEncoder(vocab_size,
num_classes=10,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
max_len=2048,
... | [
"def",
"TransformerEncoder",
"(",
"vocab_size",
",",
"num_classes",
"=",
"10",
",",
"feature_depth",
"=",
"512",
",",
"feedforward_depth",
"=",
"2048",
",",
"num_layers",
"=",
"6",
",",
"num_heads",
"=",
"8",
",",
"dropout",
"=",
"0.1",
",",
"max_len",
"="... | Transformer encoder.
Args:
vocab_size: int: vocab size
num_classes: how many classes on output
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_layers: int: number of encoder/decoder layers
num_heads: int: number of attention heads
dropout: f... | [
"Transformer",
"encoder",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L79-L120 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | DecoderLayer | def DecoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
mode):
"""Transformer decoder layer.
Args:
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_heads: int: number of att... | python | def DecoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
mode):
"""Transformer decoder layer.
Args:
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_heads: int: number of att... | [
"def",
"DecoderLayer",
"(",
"feature_depth",
",",
"feedforward_depth",
",",
"num_heads",
",",
"dropout",
",",
"mode",
")",
":",
"return",
"layers",
".",
"Serial",
"(",
"layers",
".",
"Residual",
"(",
"# Self-attention block.",
"layers",
".",
"LayerNorm",
"(",
... | Transformer decoder layer.
Args:
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_heads: int: number of attention heads
dropout: float: dropout rate (how much to drop out)
mode: str: 'train' or 'eval'
Returns:
the layer. | [
"Transformer",
"decoder",
"layer",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L123-L151 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | TransformerLM | def TransformerLM(vocab_size,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
max_len=2048,
mode='train'):
"""Transformer language model (only uses the decod... | python | def TransformerLM(vocab_size,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
max_len=2048,
mode='train'):
"""Transformer language model (only uses the decod... | [
"def",
"TransformerLM",
"(",
"vocab_size",
",",
"feature_depth",
"=",
"512",
",",
"feedforward_depth",
"=",
"2048",
",",
"num_layers",
"=",
"6",
",",
"num_heads",
"=",
"8",
",",
"dropout",
"=",
"0.1",
",",
"max_len",
"=",
"2048",
",",
"mode",
"=",
"'trai... | Transformer language model (only uses the decoder part of Transformer).
Args:
vocab_size: int: vocab size
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_layers: int: number of encoder/decoder layers
num_heads: int: number of attention heads
dro... | [
"Transformer",
"language",
"model",
"(",
"only",
"uses",
"the",
"decoder",
"part",
"of",
"Transformer",
")",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L154-L188 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | ChunkedDecoderLayer | def ChunkedDecoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
chunk_selector,
mode):
"""Transformer decoder layer operating on chunks.
Args:
feature_depth: int: depth of embe... | python | def ChunkedDecoderLayer(feature_depth,
feedforward_depth,
num_heads,
dropout,
chunk_selector,
mode):
"""Transformer decoder layer operating on chunks.
Args:
feature_depth: int: depth of embe... | [
"def",
"ChunkedDecoderLayer",
"(",
"feature_depth",
",",
"feedforward_depth",
",",
"num_heads",
",",
"dropout",
",",
"chunk_selector",
",",
"mode",
")",
":",
"return",
"layers",
".",
"Serial",
"(",
"layers",
".",
"Residual",
"(",
"# Self-attention block.",
"layers... | Transformer decoder layer operating on chunks.
Args:
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_heads: int: number of attention heads
dropout: float: dropout rate (how much to drop out)
chunk_selector: a function from chunk number to list of ch... | [
"Transformer",
"decoder",
"layer",
"operating",
"on",
"chunks",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L191-L220 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | ChunkedTransformerLM | def ChunkedTransformerLM(vocab_size,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
chunk_selector=None,
max_... | python | def ChunkedTransformerLM(vocab_size,
feature_depth=512,
feedforward_depth=2048,
num_layers=6,
num_heads=8,
dropout=0.1,
chunk_selector=None,
max_... | [
"def",
"ChunkedTransformerLM",
"(",
"vocab_size",
",",
"feature_depth",
"=",
"512",
",",
"feedforward_depth",
"=",
"2048",
",",
"num_layers",
"=",
"6",
",",
"num_heads",
"=",
"8",
",",
"dropout",
"=",
"0.1",
",",
"chunk_selector",
"=",
"None",
",",
"max_len"... | Transformer language model operating on chunks.
The input to this model is a sequence presented as a list or tuple of chunks:
(chunk1, chunk2, chunks3, ..., chunkN).
Each chunk should have the same shape (batch, chunk-length) and together they
represent a long sequence that's a concatenation chunk1,chunk2,.... | [
"Transformer",
"language",
"model",
"operating",
"on",
"chunks",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L223-L273 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/transformer.py | Transformer | def Transformer(source_vocab_size,
target_vocab_size,
mode='train',
num_layers=6,
feature_depth=512,
feedforward_depth=2048,
num_heads=8,
dropout=0.1,
shared_embedding=True,
ma... | python | def Transformer(source_vocab_size,
target_vocab_size,
mode='train',
num_layers=6,
feature_depth=512,
feedforward_depth=2048,
num_heads=8,
dropout=0.1,
shared_embedding=True,
ma... | [
"def",
"Transformer",
"(",
"source_vocab_size",
",",
"target_vocab_size",
",",
"mode",
"=",
"'train'",
",",
"num_layers",
"=",
"6",
",",
"feature_depth",
"=",
"512",
",",
"feedforward_depth",
"=",
"2048",
",",
"num_heads",
"=",
"8",
",",
"dropout",
"=",
"0.1... | Transformer model.
Args:
source_vocab_size: int: source vocab size
target_vocab_size: int: target vocab size
mode: str: 'train' or 'eval'
num_layers: int: number of encoder/decoder layers
feature_depth: int: depth of embedding
feedforward_depth: int: depth of feed-forward layer
num_heads... | [
"Transformer",
"model",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L279-L431 | train |
tensorflow/tensor2tensor | tensor2tensor/models/mtf_transformer.py | mtf_transformer_base | def mtf_transformer_base():
"""Set of hyperparameters."""
hparams = common_hparams.basic_params1()
hparams.no_data_parallelism = True
hparams.use_fixed_batch_size = True
hparams.add_hparam("mtf_mode", True)
hparams.batch_size = 64
hparams.max_length = 256
hparams.add_hparam("d_model", 512)
hparams.add... | python | def mtf_transformer_base():
"""Set of hyperparameters."""
hparams = common_hparams.basic_params1()
hparams.no_data_parallelism = True
hparams.use_fixed_batch_size = True
hparams.add_hparam("mtf_mode", True)
hparams.batch_size = 64
hparams.max_length = 256
hparams.add_hparam("d_model", 512)
hparams.add... | [
"def",
"mtf_transformer_base",
"(",
")",
":",
"hparams",
"=",
"common_hparams",
".",
"basic_params1",
"(",
")",
"hparams",
".",
"no_data_parallelism",
"=",
"True",
"hparams",
".",
"use_fixed_batch_size",
"=",
"True",
"hparams",
".",
"add_hparam",
"(",
"\"mtf_mode\... | Set of hyperparameters. | [
"Set",
"of",
"hyperparameters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L791-L883 | train |
tensorflow/tensor2tensor | tensor2tensor/models/mtf_transformer.py | mtf_transformer_tiny | def mtf_transformer_tiny():
"""Catch bugs locally..."""
hparams = mtf_transformer_base()
hparams.d_model = 128
hparams.d_ff = 512
hparams.batch_size = 8
hparams.encoder_layers = ["att", "drd"] * 2
hparams.decoder_layers = ["att", "enc_att", "drd"] * 2
hparams.num_heads = 8
# data parallelism and model... | python | def mtf_transformer_tiny():
"""Catch bugs locally..."""
hparams = mtf_transformer_base()
hparams.d_model = 128
hparams.d_ff = 512
hparams.batch_size = 8
hparams.encoder_layers = ["att", "drd"] * 2
hparams.decoder_layers = ["att", "enc_att", "drd"] * 2
hparams.num_heads = 8
# data parallelism and model... | [
"def",
"mtf_transformer_tiny",
"(",
")",
":",
"hparams",
"=",
"mtf_transformer_base",
"(",
")",
"hparams",
".",
"d_model",
"=",
"128",
"hparams",
".",
"d_ff",
"=",
"512",
"hparams",
".",
"batch_size",
"=",
"8",
"hparams",
".",
"encoder_layers",
"=",
"[",
"... | Catch bugs locally... | [
"Catch",
"bugs",
"locally",
"..."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L897-L909 | train |
tensorflow/tensor2tensor | tensor2tensor/models/mtf_transformer.py | mtf_transformer_paper_lm | def mtf_transformer_paper_lm(size):
"""Config for language-model experiments.
Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles e... | python | def mtf_transformer_paper_lm(size):
"""Config for language-model experiments.
Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles e... | [
"def",
"mtf_transformer_paper_lm",
"(",
"size",
")",
":",
"n",
"=",
"2",
"**",
"size",
"hparams",
"=",
"mtf_transformer_base_lm",
"(",
")",
"hparams",
".",
"batch_size",
"=",
"256",
"hparams",
".",
"d_model",
"=",
"1024",
"hparams",
".",
"d_ff",
"=",
"int"... | Config for language-model experiments.
Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles each of these.
Results:
size params... | [
"Config",
"for",
"language",
"-",
"model",
"experiments",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L953-L989 | train |
tensorflow/tensor2tensor | tensor2tensor/models/mtf_transformer.py | mtf_transformer_paper_tr | def mtf_transformer_paper_tr(size):
"""Config for translation experiments.
Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles each ... | python | def mtf_transformer_paper_tr(size):
"""Config for translation experiments.
Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles each ... | [
"def",
"mtf_transformer_paper_tr",
"(",
"size",
")",
":",
"n",
"=",
"2",
"**",
"size",
"hparams",
"=",
"mtf_transformer_base",
"(",
")",
"hparams",
".",
"label_smoothing",
"=",
"0.1",
"hparams",
".",
"batch_size",
"=",
"128",
"hparams",
".",
"d_model",
"=",
... | Config for translation experiments.
Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs)
The size parameter is an integer that controls the number of heads and the
size of the size of the feedforward hidden layers. Increasing size by 1
doubles each of these.
Args:
size: an integer
... | [
"Config",
"for",
"translation",
"experiments",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1041-L1065 | train |
tensorflow/tensor2tensor | tensor2tensor/models/mtf_transformer.py | mtf_transformer_lm_baseline | def mtf_transformer_lm_baseline():
"""Small language model to run on 1 TPU.
Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs)
Results:
params/10^9 log-ppl(per-token)
0.14 3.202
Returns:
a hparams
"""
hparams = mtf_transformer_paper_lm(-1)
hparams... | python | def mtf_transformer_lm_baseline():
"""Small language model to run on 1 TPU.
Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs)
Results:
params/10^9 log-ppl(per-token)
0.14 3.202
Returns:
a hparams
"""
hparams = mtf_transformer_paper_lm(-1)
hparams... | [
"def",
"mtf_transformer_lm_baseline",
"(",
")",
":",
"hparams",
"=",
"mtf_transformer_paper_lm",
"(",
"-",
"1",
")",
"hparams",
".",
"batch_size",
"=",
"128",
"hparams",
".",
"learning_rate_decay_steps",
"=",
"27200",
"# one epoch on lm1b",
"hparams",
".",
"mesh_sha... | Small language model to run on 1 TPU.
Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs)
Results:
params/10^9 log-ppl(per-token)
0.14 3.202
Returns:
a hparams | [
"Small",
"language",
"model",
"to",
"run",
"on",
"1",
"TPU",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1171-L1186 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | multihead_graph_attention | def multihead_graph_attention(query_antecedent,
memory_antecedent,
bias,
total_key_depth,
total_value_depth,
output_depth,
num_heads,
... | python | def multihead_graph_attention(query_antecedent,
memory_antecedent,
bias,
total_key_depth,
total_value_depth,
output_depth,
num_heads,
... | [
"def",
"multihead_graph_attention",
"(",
"query_antecedent",
",",
"memory_antecedent",
",",
"bias",
",",
"total_key_depth",
",",
"total_value_depth",
",",
"output_depth",
",",
"num_heads",
",",
"dropout_rate",
",",
"image_shapes",
"=",
"None",
",",
"attention_type",
"... | Multihead scaled-dot-product attention with input/output transformations.
Args:
query_antecedent: a Tensor with shape [batch, length_q, channels]
memory_antecedent: a Tensor with shape [batch, length_m, channels] or None
bias: bias Tensor (see attention_bias())
total_key_depth: an integer
total_v... | [
"Multihead",
"scaled",
"-",
"dot",
"-",
"product",
"attention",
"with",
"input",
"/",
"output",
"transformations",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L28-L146 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | graph_attention | def graph_attention(q,
k,
v,
bias,
dropout_rate=0.0,
image_shapes=None,
name=None,
make_image_summary=True,
save_weights_to=None,
dropout_br... | python | def graph_attention(q,
k,
v,
bias,
dropout_rate=0.0,
image_shapes=None,
name=None,
make_image_summary=True,
save_weights_to=None,
dropout_br... | [
"def",
"graph_attention",
"(",
"q",
",",
"k",
",",
"v",
",",
"bias",
",",
"dropout_rate",
"=",
"0.0",
",",
"image_shapes",
"=",
"None",
",",
"name",
"=",
"None",
",",
"make_image_summary",
"=",
"True",
",",
"save_weights_to",
"=",
"None",
",",
"dropout_b... | graph attention.
Args:
q: a Tensor with shape [batch, heads, length_q, depth_k]
k: a Tensor with shape [batch, heads, length_kv, depth_k]
v: a Tensor with shape [batch, heads, length_kv, depth_v]
bias: bias Tensor (see attention_bias())
dropout_rate: a floating point number
image_shapes: opti... | [
"graph",
"attention",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L185-L248 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | _compute_edge_transforms | def _compute_edge_transforms(node_states,
depth,
num_transforms,
name="transform"):
"""Helper function that computes transformation for keys and values.
Let B be the number of batches.
Let N be the number of nodes in the graph... | python | def _compute_edge_transforms(node_states,
depth,
num_transforms,
name="transform"):
"""Helper function that computes transformation for keys and values.
Let B be the number of batches.
Let N be the number of nodes in the graph... | [
"def",
"_compute_edge_transforms",
"(",
"node_states",
",",
"depth",
",",
"num_transforms",
",",
"name",
"=",
"\"transform\"",
")",
":",
"node_shapes",
"=",
"common_layers",
".",
"shape_list",
"(",
"node_states",
")",
"x",
"=",
"common_layers",
".",
"dense",
"("... | Helper function that computes transformation for keys and values.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let D be the size of the node hidden states.
Let K be the size of the attention keys/queries (total_key_depth).
Let V be the size of the attention values (total_value_d... | [
"Helper",
"function",
"that",
"computes",
"transformation",
"for",
"keys",
"and",
"values",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L251-L301 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | compute_mpnn_qkv | def compute_mpnn_qkv(node_states,
total_key_depth,
total_value_depth,
num_transforms):
"""Computes query, key and value for edge matrices.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let D be the size of the node hidd... | python | def compute_mpnn_qkv(node_states,
total_key_depth,
total_value_depth,
num_transforms):
"""Computes query, key and value for edge matrices.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let D be the size of the node hidd... | [
"def",
"compute_mpnn_qkv",
"(",
"node_states",
",",
"total_key_depth",
",",
"total_value_depth",
",",
"num_transforms",
")",
":",
"# node_states is initially a tensor with shape [B, N, D]. The call to dense",
"# creates a D x K kernel that serves as a fully-connected layer.",
"#",
"# F... | Computes query, key and value for edge matrices.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let D be the size of the node hidden states.
Let K be the size of the attention keys/queries (total_key_depth).
Let V be the size of the attention values (total_value_depth).
Let T be... | [
"Computes",
"query",
"key",
"and",
"value",
"for",
"edge",
"matrices",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L304-L364 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | sparse_message_pass_batched | def sparse_message_pass_batched(node_states,
adjacency_matrices,
num_edge_types,
hidden_size,
use_bias=True,
average_aggregation=False,
... | python | def sparse_message_pass_batched(node_states,
adjacency_matrices,
num_edge_types,
hidden_size,
use_bias=True,
average_aggregation=False,
... | [
"def",
"sparse_message_pass_batched",
"(",
"node_states",
",",
"adjacency_matrices",
",",
"num_edge_types",
",",
"hidden_size",
",",
"use_bias",
"=",
"True",
",",
"average_aggregation",
"=",
"False",
",",
"name",
"=",
"\"sparse_ggnn_batched\"",
")",
":",
"b",
",",
... | Identical to sparse_ggnn except that each input has a batch dimension.
B = The batch size.
N = The number of nodes in each batch.
H = The size of the hidden states.
T = The number of edge types.
Args:
node_states: Initial states of each node in the graph. Shape: [B, N, H]
adjacency_matrices: Adjacen... | [
"Identical",
"to",
"sparse_ggnn",
"except",
"that",
"each",
"input",
"has",
"a",
"batch",
"dimension",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L367-L428 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | sparse_message_pass | def sparse_message_pass(node_states,
adjacency_matrices,
num_edge_types,
hidden_size,
use_bias=True,
average_aggregation=False,
name="sparse_ggnn"):
"""One message-passing st... | python | def sparse_message_pass(node_states,
adjacency_matrices,
num_edge_types,
hidden_size,
use_bias=True,
average_aggregation=False,
name="sparse_ggnn"):
"""One message-passing st... | [
"def",
"sparse_message_pass",
"(",
"node_states",
",",
"adjacency_matrices",
",",
"num_edge_types",
",",
"hidden_size",
",",
"use_bias",
"=",
"True",
",",
"average_aggregation",
"=",
"False",
",",
"name",
"=",
"\"sparse_ggnn\"",
")",
":",
"n",
"=",
"tf",
".",
... | One message-passing step for a GNN with a sparse adjacency matrix.
Implements equation 2 (the message passing step) in
[Li et al. 2015](https://arxiv.org/abs/1511.05493).
N = The number of nodes in each batch.
H = The size of the hidden states.
T = The number of edge types.
Args:
node_states: Initial... | [
"One",
"message",
"-",
"passing",
"step",
"for",
"a",
"GNN",
"with",
"a",
"sparse",
"adjacency",
"matrix",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L431-L511 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | multihead_mpnn_attention | def multihead_mpnn_attention(node_states,
total_key_depth,
total_value_depth,
output_depth,
num_heads,
adjacency_matrix=None,
num_edge_types=5,
... | python | def multihead_mpnn_attention(node_states,
total_key_depth,
total_value_depth,
output_depth,
num_heads,
adjacency_matrix=None,
num_edge_types=5,
... | [
"def",
"multihead_mpnn_attention",
"(",
"node_states",
",",
"total_key_depth",
",",
"total_value_depth",
",",
"output_depth",
",",
"num_heads",
",",
"adjacency_matrix",
"=",
"None",
",",
"num_edge_types",
"=",
"5",
",",
"num_transforms",
"=",
"None",
",",
"use_weigh... | Multihead scaled-dot-product attention with input/output transformations.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let D be the size of the node hidden states.
Let K be the size of the attention keys/queries (total_key_depth).
Let V be the size of the attention values (total... | [
"Multihead",
"scaled",
"-",
"dot",
"-",
"product",
"attention",
"with",
"input",
"/",
"output",
"transformations",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L514-L649 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | dot_product_mpnn_attention | def dot_product_mpnn_attention(q,
k,
v,
adjacency_matrix,
num_edge_types,
num_transforms=None,
use_weighted_sum=False,
... | python | def dot_product_mpnn_attention(q,
k,
v,
adjacency_matrix,
num_edge_types,
num_transforms=None,
use_weighted_sum=False,
... | [
"def",
"dot_product_mpnn_attention",
"(",
"q",
",",
"k",
",",
"v",
",",
"adjacency_matrix",
",",
"num_edge_types",
",",
"num_transforms",
"=",
"None",
",",
"use_weighted_sum",
"=",
"False",
",",
"name",
"=",
"None",
")",
":",
"with",
"tf",
".",
"variable_sco... | Dot product attention with edge vectors.
Let B be the number of batches.
Let N be the number of nodes in the graph.
Let K be the size of the attention keys/queries.
Let V be the size of the attention values.
Let T be the total number of transforms (num_transforms).
Args:
q: The query Tensor of shape [... | [
"Dot",
"product",
"attention",
"with",
"edge",
"vectors",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L652-L783 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | ggnn_fast_dense | def ggnn_fast_dense(node_states,
adjacency_matrix,
num_edge_types,
total_value_depth,
name=None):
"""ggnn version of the MPNN from Gilmer et al.
Let B be the number of batches.
Let D be the size of the node hidden states.
Let K be ... | python | def ggnn_fast_dense(node_states,
adjacency_matrix,
num_edge_types,
total_value_depth,
name=None):
"""ggnn version of the MPNN from Gilmer et al.
Let B be the number of batches.
Let D be the size of the node hidden states.
Let K be ... | [
"def",
"ggnn_fast_dense",
"(",
"node_states",
",",
"adjacency_matrix",
",",
"num_edge_types",
",",
"total_value_depth",
",",
"name",
"=",
"None",
")",
":",
"# between the same nodes (with only one edge of each type. adjacency_matrix",
"# will need to be converted to shape [B, T, N,... | ggnn version of the MPNN from Gilmer et al.
Let B be the number of batches.
Let D be the size of the node hidden states.
Let K be the size of the attention keys/queries.
Let V be the size of the output of the ggnn.
Let T be the number of transforms / edge types.
Args:
node_states: The value Tensor of ... | [
"ggnn",
"version",
"of",
"the",
"MPNN",
"from",
"Gilmer",
"et",
"al",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L786-L837 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | compute_values | def compute_values(edge_compatibility, v):
"""Compute values. If edge compatibilities is just adjacency, we get ggnn.
Args:
edge_compatibility: A tensor of shape [batch, num_transforms, length, depth]
v: A tensor of shape [batch, num_transforms, length, depth]
Returns:
output: A [batch, length, dept... | python | def compute_values(edge_compatibility, v):
"""Compute values. If edge compatibilities is just adjacency, we get ggnn.
Args:
edge_compatibility: A tensor of shape [batch, num_transforms, length, depth]
v: A tensor of shape [batch, num_transforms, length, depth]
Returns:
output: A [batch, length, dept... | [
"def",
"compute_values",
"(",
"edge_compatibility",
",",
"v",
")",
":",
"# Computes the incoming value vectors for each node by weighting them",
"# according to the attention weights. These values are still segregated by",
"# edge type.",
"# Shape = [B, T, N, V].",
"all_edge_values",
"=",
... | Compute values. If edge compatibilities is just adjacency, we get ggnn.
Args:
edge_compatibility: A tensor of shape [batch, num_transforms, length, depth]
v: A tensor of shape [batch, num_transforms, length, depth]
Returns:
output: A [batch, length, depth] tensor | [
"Compute",
"values",
".",
"If",
"edge",
"compatibilities",
"is",
"just",
"adjacency",
"we",
"get",
"ggnn",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L840-L860 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | precompute_edge_matrices | def precompute_edge_matrices(adjacency, hparams):
"""Precompute the a_in and a_out tensors.
(we don't want to add to the graph everytime _fprop is called)
Args:
adjacency: placeholder of real valued vectors of shape [B, L, L, E]
hparams: HParams object
Returns:
edge_matrices: [batch, L * D, L * D] ... | python | def precompute_edge_matrices(adjacency, hparams):
"""Precompute the a_in and a_out tensors.
(we don't want to add to the graph everytime _fprop is called)
Args:
adjacency: placeholder of real valued vectors of shape [B, L, L, E]
hparams: HParams object
Returns:
edge_matrices: [batch, L * D, L * D] ... | [
"def",
"precompute_edge_matrices",
"(",
"adjacency",
",",
"hparams",
")",
":",
"batch_size",
",",
"num_nodes",
",",
"_",
",",
"edge_dim",
"=",
"common_layers",
".",
"shape_list",
"(",
"adjacency",
")",
"# build the edge_network for incoming edges",
"with",
"tf",
"."... | Precompute the a_in and a_out tensors.
(we don't want to add to the graph everytime _fprop is called)
Args:
adjacency: placeholder of real valued vectors of shape [B, L, L, E]
hparams: HParams object
Returns:
edge_matrices: [batch, L * D, L * D] the dense matrix for message passing
viewed as a bl... | [
"Precompute",
"the",
"a_in",
"and",
"a_out",
"tensors",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L863-L907 | train |
tensorflow/tensor2tensor | tensor2tensor/layers/message_passing_attention.py | dense_message_pass | def dense_message_pass(node_states, edge_matrices):
"""Computes a_t from h_{t-1}, see bottom of page 3 in the paper.
Args:
node_states: [B, L, D] tensor (h_{t-1})
edge_matrices (tf.float32): [B, L*D, L*D]
Returns:
messages (tf.float32): [B, L, D] For each pair
of nodes in the graph a message i... | python | def dense_message_pass(node_states, edge_matrices):
"""Computes a_t from h_{t-1}, see bottom of page 3 in the paper.
Args:
node_states: [B, L, D] tensor (h_{t-1})
edge_matrices (tf.float32): [B, L*D, L*D]
Returns:
messages (tf.float32): [B, L, D] For each pair
of nodes in the graph a message i... | [
"def",
"dense_message_pass",
"(",
"node_states",
",",
"edge_matrices",
")",
":",
"batch_size",
",",
"num_nodes",
",",
"node_dim",
"=",
"common_layers",
".",
"shape_list",
"(",
"node_states",
")",
"# Stack the nodes as a big column vector.",
"h_flat",
"=",
"tf",
".",
... | Computes a_t from h_{t-1}, see bottom of page 3 in the paper.
Args:
node_states: [B, L, D] tensor (h_{t-1})
edge_matrices (tf.float32): [B, L*D, L*D]
Returns:
messages (tf.float32): [B, L, D] For each pair
of nodes in the graph a message is sent along both the incoming and
outgoing edge. | [
"Computes",
"a_t",
"from",
"h_",
"{",
"t",
"-",
"1",
"}",
"see",
"bottom",
"of",
"page",
"3",
"in",
"the",
"paper",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L910-L935 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | to_example | def to_example(dictionary):
"""Helper: build tf.Example from (string -> int/float/str list) dictionary."""
features = {}
for (k, v) in six.iteritems(dictionary):
if not v:
raise ValueError("Empty generated field: %s" % str((k, v)))
if isinstance(v[0], six.integer_types):
features[k] = tf.train... | python | def to_example(dictionary):
"""Helper: build tf.Example from (string -> int/float/str list) dictionary."""
features = {}
for (k, v) in six.iteritems(dictionary):
if not v:
raise ValueError("Empty generated field: %s" % str((k, v)))
if isinstance(v[0], six.integer_types):
features[k] = tf.train... | [
"def",
"to_example",
"(",
"dictionary",
")",
":",
"features",
"=",
"{",
"}",
"for",
"(",
"k",
",",
"v",
")",
"in",
"six",
".",
"iteritems",
"(",
"dictionary",
")",
":",
"if",
"not",
"v",
":",
"raise",
"ValueError",
"(",
"\"Empty generated field: %s\"",
... | Helper: build tf.Example from (string -> int/float/str list) dictionary. | [
"Helper",
":",
"build",
"tf",
".",
"Example",
"from",
"(",
"string",
"-",
">",
"int",
"/",
"float",
"/",
"str",
"list",
")",
"dictionary",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L43-L62 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | generate_files_distributed | def generate_files_distributed(generator,
output_name,
output_dir,
num_shards=1,
max_cases=None,
task_id=0):
"""generate_files but with a single writer writing to ... | python | def generate_files_distributed(generator,
output_name,
output_dir,
num_shards=1,
max_cases=None,
task_id=0):
"""generate_files but with a single writer writing to ... | [
"def",
"generate_files_distributed",
"(",
"generator",
",",
"output_name",
",",
"output_dir",
",",
"num_shards",
"=",
"1",
",",
"max_cases",
"=",
"None",
",",
"task_id",
"=",
"0",
")",
":",
"assert",
"task_id",
"<",
"num_shards",
"output_filename",
"=",
"shard... | generate_files but with a single writer writing to shard task_id. | [
"generate_files",
"but",
"with",
"a",
"single",
"writer",
"writing",
"to",
"shard",
"task_id",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L65-L89 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | generate_files | def generate_files(generator, output_filenames,
max_cases=None, cycle_every_n=1):
"""Generate cases from a generator and save as TFRecord files.
Generated cases are transformed to tf.Example protos and saved as TFRecords
in sharded files named output_dir/output_name-00..N-of-00..M=num_shards.
... | python | def generate_files(generator, output_filenames,
max_cases=None, cycle_every_n=1):
"""Generate cases from a generator and save as TFRecord files.
Generated cases are transformed to tf.Example protos and saved as TFRecords
in sharded files named output_dir/output_name-00..N-of-00..M=num_shards.
... | [
"def",
"generate_files",
"(",
"generator",
",",
"output_filenames",
",",
"max_cases",
"=",
"None",
",",
"cycle_every_n",
"=",
"1",
")",
":",
"if",
"outputs_exist",
"(",
"output_filenames",
")",
":",
"tf",
".",
"logging",
".",
"info",
"(",
"\"Skipping generator... | Generate cases from a generator and save as TFRecord files.
Generated cases are transformed to tf.Example protos and saved as TFRecords
in sharded files named output_dir/output_name-00..N-of-00..M=num_shards.
Args:
generator: a generator yielding (string -> int/float/str list) dictionaries.
output_filen... | [
"Generate",
"cases",
"from",
"a",
"generator",
"and",
"save",
"as",
"TFRecord",
"files",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L134-L193 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | download_report_hook | def download_report_hook(count, block_size, total_size):
"""Report hook for download progress.
Args:
count: current block number
block_size: block size
total_size: total size
"""
percent = int(count * block_size * 100 / total_size)
print("\r%d%%" % percent + " completed", end="\r") | python | def download_report_hook(count, block_size, total_size):
"""Report hook for download progress.
Args:
count: current block number
block_size: block size
total_size: total size
"""
percent = int(count * block_size * 100 / total_size)
print("\r%d%%" % percent + " completed", end="\r") | [
"def",
"download_report_hook",
"(",
"count",
",",
"block_size",
",",
"total_size",
")",
":",
"percent",
"=",
"int",
"(",
"count",
"*",
"block_size",
"*",
"100",
"/",
"total_size",
")",
"print",
"(",
"\"\\r%d%%\"",
"%",
"percent",
"+",
"\" completed\"",
",",
... | Report hook for download progress.
Args:
count: current block number
block_size: block size
total_size: total size | [
"Report",
"hook",
"for",
"download",
"progress",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L196-L205 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | maybe_download | def maybe_download(directory, filename, uri):
"""Download filename from uri unless it's already in directory.
Copies a remote file to local if that local file does not already exist. If
the local file pre-exists this function call, it does not check that the local
file is a copy of the remote.
Remote filen... | python | def maybe_download(directory, filename, uri):
"""Download filename from uri unless it's already in directory.
Copies a remote file to local if that local file does not already exist. If
the local file pre-exists this function call, it does not check that the local
file is a copy of the remote.
Remote filen... | [
"def",
"maybe_download",
"(",
"directory",
",",
"filename",
",",
"uri",
")",
":",
"tf",
".",
"gfile",
".",
"MakeDirs",
"(",
"directory",
")",
"filepath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"directory",
",",
"filename",
")",
"if",
"tf",
".",
"g... | Download filename from uri unless it's already in directory.
Copies a remote file to local if that local file does not already exist. If
the local file pre-exists this function call, it does not check that the local
file is a copy of the remote.
Remote filenames can be filepaths, any URI readable by tensorfl... | [
"Download",
"filename",
"from",
"uri",
"unless",
"it",
"s",
"already",
"in",
"directory",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L208-L248 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | maybe_download_from_drive | def maybe_download_from_drive(directory, filename, url):
"""Download filename from Google drive unless it's already in directory.
Args:
directory: path to the directory that will be used.
filename: name of the file to download to (do nothing if it already exists).
url: URL to download from.
Returns:... | python | def maybe_download_from_drive(directory, filename, url):
"""Download filename from Google drive unless it's already in directory.
Args:
directory: path to the directory that will be used.
filename: name of the file to download to (do nothing if it already exists).
url: URL to download from.
Returns:... | [
"def",
"maybe_download_from_drive",
"(",
"directory",
",",
"filename",
",",
"url",
")",
":",
"if",
"not",
"tf",
".",
"gfile",
".",
"Exists",
"(",
"directory",
")",
":",
"tf",
".",
"logging",
".",
"info",
"(",
"\"Creating directory %s\"",
"%",
"directory",
... | Download filename from Google drive unless it's already in directory.
Args:
directory: path to the directory that will be used.
filename: name of the file to download to (do nothing if it already exists).
url: URL to download from.
Returns:
The path to the downloaded file. | [
"Download",
"filename",
"from",
"Google",
"drive",
"unless",
"it",
"s",
"already",
"in",
"directory",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L251-L298 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | gunzip_file | def gunzip_file(gz_path, new_path):
"""Unzips from gz_path into new_path.
Args:
gz_path: path to the zipped file.
new_path: path to where the file will be unzipped.
"""
if tf.gfile.Exists(new_path):
tf.logging.info("File %s already exists, skipping unpacking" % new_path)
return
tf.logging.inf... | python | def gunzip_file(gz_path, new_path):
"""Unzips from gz_path into new_path.
Args:
gz_path: path to the zipped file.
new_path: path to where the file will be unzipped.
"""
if tf.gfile.Exists(new_path):
tf.logging.info("File %s already exists, skipping unpacking" % new_path)
return
tf.logging.inf... | [
"def",
"gunzip_file",
"(",
"gz_path",
",",
"new_path",
")",
":",
"if",
"tf",
".",
"gfile",
".",
"Exists",
"(",
"new_path",
")",
":",
"tf",
".",
"logging",
".",
"info",
"(",
"\"File %s already exists, skipping unpacking\"",
"%",
"new_path",
")",
"return",
"tf... | Unzips from gz_path into new_path.
Args:
gz_path: path to the zipped file.
new_path: path to where the file will be unzipped. | [
"Unzips",
"from",
"gz_path",
"into",
"new_path",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L301-L318 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | get_or_generate_vocab_inner | def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size,
generator, max_subtoken_length=None,
reserved_tokens=None):
"""Inner implementation for vocab generators.
Args:
data_dir: The base directory where data and vocab files are store... | python | def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size,
generator, max_subtoken_length=None,
reserved_tokens=None):
"""Inner implementation for vocab generators.
Args:
data_dir: The base directory where data and vocab files are store... | [
"def",
"get_or_generate_vocab_inner",
"(",
"data_dir",
",",
"vocab_filename",
",",
"vocab_size",
",",
"generator",
",",
"max_subtoken_length",
"=",
"None",
",",
"reserved_tokens",
"=",
"None",
")",
":",
"if",
"data_dir",
"and",
"vocab_filename",
":",
"vocab_filepath... | Inner implementation for vocab generators.
Args:
data_dir: The base directory where data and vocab files are stored. If None,
then do not save the vocab even if it doesn't exist.
vocab_filename: relative filename where vocab file is stored
vocab_size: target size of the vocabulary constructed by Su... | [
"Inner",
"implementation",
"for",
"vocab",
"generators",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L321-L358 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | get_or_generate_vocab | def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size,
sources, file_byte_budget=1e6,
max_subtoken_length=None):
"""Generate a vocabulary from the datasets in sources."""
vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_bud... | python | def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size,
sources, file_byte_budget=1e6,
max_subtoken_length=None):
"""Generate a vocabulary from the datasets in sources."""
vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_bud... | [
"def",
"get_or_generate_vocab",
"(",
"data_dir",
",",
"tmp_dir",
",",
"vocab_filename",
",",
"vocab_size",
",",
"sources",
",",
"file_byte_budget",
"=",
"1e6",
",",
"max_subtoken_length",
"=",
"None",
")",
":",
"vocab_generator",
"=",
"generate_lines_for_vocab",
"("... | Generate a vocabulary from the datasets in sources. | [
"Generate",
"a",
"vocabulary",
"from",
"the",
"datasets",
"in",
"sources",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L361-L368 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | generate_lines_for_vocab | def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6):
"""Generate lines for vocabulary generation."""
tf.logging.info("Generating vocab from: %s", str(sources))
for source in sources:
url = source[0]
filename = os.path.basename(url)
compressed_file = maybe_download(tmp_dir, filename, u... | python | def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6):
"""Generate lines for vocabulary generation."""
tf.logging.info("Generating vocab from: %s", str(sources))
for source in sources:
url = source[0]
filename = os.path.basename(url)
compressed_file = maybe_download(tmp_dir, filename, u... | [
"def",
"generate_lines_for_vocab",
"(",
"tmp_dir",
",",
"sources",
",",
"file_byte_budget",
"=",
"1e6",
")",
":",
"tf",
".",
"logging",
".",
"info",
"(",
"\"Generating vocab from: %s\"",
",",
"str",
"(",
"sources",
")",
")",
"for",
"source",
"in",
"sources",
... | Generate lines for vocabulary generation. | [
"Generate",
"lines",
"for",
"vocabulary",
"generation",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L371-L413 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | get_or_generate_tabbed_vocab | def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename,
index, vocab_filename, vocab_size):
r"""Generate a vocabulary from a tabbed source file.
The source is a file of source, target pairs, where each line contains
a source string and a target string, separated by a... | python | def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename,
index, vocab_filename, vocab_size):
r"""Generate a vocabulary from a tabbed source file.
The source is a file of source, target pairs, where each line contains
a source string and a target string, separated by a... | [
"def",
"get_or_generate_tabbed_vocab",
"(",
"data_dir",
",",
"tmp_dir",
",",
"source_filename",
",",
"index",
",",
"vocab_filename",
",",
"vocab_size",
")",
":",
"def",
"generate",
"(",
")",
":",
"filepath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"tmp_dir... | r"""Generate a vocabulary from a tabbed source file.
The source is a file of source, target pairs, where each line contains
a source string and a target string, separated by a tab ('\t') character.
The index parameter specifies 0 for the source or 1 for the target.
Args:
data_dir: path to the data directo... | [
"r",
"Generate",
"a",
"vocabulary",
"from",
"a",
"tabbed",
"source",
"file",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L416-L447 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | get_or_generate_txt_vocab | def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size,
filepatterns):
"""Generate a vocabulary from txt files with example-per-line."""
if isinstance(filepatterns, str):
filepatterns = [filepatterns]
def generate():
tf.logging.info("Generating vocab from %s", fi... | python | def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size,
filepatterns):
"""Generate a vocabulary from txt files with example-per-line."""
if isinstance(filepatterns, str):
filepatterns = [filepatterns]
def generate():
tf.logging.info("Generating vocab from %s", fi... | [
"def",
"get_or_generate_txt_vocab",
"(",
"data_dir",
",",
"vocab_filename",
",",
"vocab_size",
",",
"filepatterns",
")",
":",
"if",
"isinstance",
"(",
"filepatterns",
",",
"str",
")",
":",
"filepatterns",
"=",
"[",
"filepatterns",
"]",
"def",
"generate",
"(",
... | Generate a vocabulary from txt files with example-per-line. | [
"Generate",
"a",
"vocabulary",
"from",
"txt",
"files",
"with",
"example",
"-",
"per",
"-",
"line",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L450-L465 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | _shuffle_single | def _shuffle_single(fname, extra_fn=None):
"""Shuffle a single file of records.
Args:
fname: a string
extra_fn: an optional function from list of TFRecords to list of TFRecords
to be called after shuffling.
"""
records = read_records(fname)
random.shuffle(records)
if extra_fn is not None:
... | python | def _shuffle_single(fname, extra_fn=None):
"""Shuffle a single file of records.
Args:
fname: a string
extra_fn: an optional function from list of TFRecords to list of TFRecords
to be called after shuffling.
"""
records = read_records(fname)
random.shuffle(records)
if extra_fn is not None:
... | [
"def",
"_shuffle_single",
"(",
"fname",
",",
"extra_fn",
"=",
"None",
")",
":",
"records",
"=",
"read_records",
"(",
"fname",
")",
"random",
".",
"shuffle",
"(",
"records",
")",
"if",
"extra_fn",
"is",
"not",
"None",
":",
"records",
"=",
"extra_fn",
"(",... | Shuffle a single file of records.
Args:
fname: a string
extra_fn: an optional function from list of TFRecords to list of TFRecords
to be called after shuffling. | [
"Shuffle",
"a",
"single",
"file",
"of",
"records",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L499-L513 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | shuffle_dataset | def shuffle_dataset(filenames, extra_fn=None):
"""Shuffles the dataset.
Args:
filenames: a list of strings
extra_fn: an optional function from list of records to list of records
to be called after shuffling a file.
"""
if outputs_exist(filenames):
tf.logging.info("Skipping shuffle because out... | python | def shuffle_dataset(filenames, extra_fn=None):
"""Shuffles the dataset.
Args:
filenames: a list of strings
extra_fn: an optional function from list of records to list of records
to be called after shuffling a file.
"""
if outputs_exist(filenames):
tf.logging.info("Skipping shuffle because out... | [
"def",
"shuffle_dataset",
"(",
"filenames",
",",
"extra_fn",
"=",
"None",
")",
":",
"if",
"outputs_exist",
"(",
"filenames",
")",
":",
"tf",
".",
"logging",
".",
"info",
"(",
"\"Skipping shuffle because output files exist\"",
")",
"return",
"tf",
".",
"logging",... | Shuffles the dataset.
Args:
filenames: a list of strings
extra_fn: an optional function from list of records to list of records
to be called after shuffling a file. | [
"Shuffles",
"the",
"dataset",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L516-L530 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | pack_examples | def pack_examples(examples,
has_inputs,
packed_length=256,
spacing=2,
queue_size=10,
chop_long_sequences=False):
"""Pack examples into longer examples.
If has_inputs=False, we are packing single-sequence examples with
targe... | python | def pack_examples(examples,
has_inputs,
packed_length=256,
spacing=2,
queue_size=10,
chop_long_sequences=False):
"""Pack examples into longer examples.
If has_inputs=False, we are packing single-sequence examples with
targe... | [
"def",
"pack_examples",
"(",
"examples",
",",
"has_inputs",
",",
"packed_length",
"=",
"256",
",",
"spacing",
"=",
"2",
",",
"queue_size",
"=",
"10",
",",
"chop_long_sequences",
"=",
"False",
")",
":",
"packer",
"=",
"SequencePairPacker",
"if",
"has_inputs",
... | Pack examples into longer examples.
If has_inputs=False, we are packing single-sequence examples with
targets only and no inputs.
In this case, we concatenate the targets from several examples to form
each new example. We insert a number of zeros for spacing between the
original sequences. This is to help... | [
"Pack",
"examples",
"into",
"longer",
"examples",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L589-L660 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | _pack_with_custom_ops | def _pack_with_custom_ops(dataset, keys, length):
"""Helper-function for packing a dataset which has already been batched.
See pack_dataset()
Relies on custom ops which require a custom compiled binary.
Faster than _pack_with_tf_ops(), and denser packing.
Args:
dataset: a dataset containing padded batc... | python | def _pack_with_custom_ops(dataset, keys, length):
"""Helper-function for packing a dataset which has already been batched.
See pack_dataset()
Relies on custom ops which require a custom compiled binary.
Faster than _pack_with_tf_ops(), and denser packing.
Args:
dataset: a dataset containing padded batc... | [
"def",
"_pack_with_custom_ops",
"(",
"dataset",
",",
"keys",
",",
"length",
")",
":",
"from",
"tensor2tensor",
".",
"data_generators",
".",
"ops",
"import",
"pack_sequences_ops",
"# pylint: disable=g-import-not-at-top",
"# faster and better packing but requires custom-built bin... | Helper-function for packing a dataset which has already been batched.
See pack_dataset()
Relies on custom ops which require a custom compiled binary.
Faster than _pack_with_tf_ops(), and denser packing.
Args:
dataset: a dataset containing padded batches of examples.
keys: a list of strings (must have... | [
"Helper",
"-",
"function",
"for",
"packing",
"a",
"dataset",
"which",
"has",
"already",
"been",
"batched",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L736-L770 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | make_tmp_dir | def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin
"""Make a temporary directory."""
if dir is None:
return tempfile.mkdtemp(suffix, prefix, dir)
else:
while True:
rand_term = random.randint(1, 9999)
tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_te... | python | def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin
"""Make a temporary directory."""
if dir is None:
return tempfile.mkdtemp(suffix, prefix, dir)
else:
while True:
rand_term = random.randint(1, 9999)
tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_te... | [
"def",
"make_tmp_dir",
"(",
"suffix",
"=",
"\"\"",
",",
"prefix",
"=",
"\"tmp\"",
",",
"dir",
"=",
"None",
")",
":",
"# pylint: disable=redefined-builtin",
"if",
"dir",
"is",
"None",
":",
"return",
"tempfile",
".",
"mkdtemp",
"(",
"suffix",
",",
"prefix",
... | Make a temporary directory. | [
"Make",
"a",
"temporary",
"directory",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L883-L895 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | tfrecord_iterator_for_problem | def tfrecord_iterator_for_problem(problem, data_dir,
dataset_split=tf.estimator.ModeKeys.TRAIN):
"""Iterate over the records on disk for the Problem."""
filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split))
example_spec = problem.example_reading_spec()[0]
... | python | def tfrecord_iterator_for_problem(problem, data_dir,
dataset_split=tf.estimator.ModeKeys.TRAIN):
"""Iterate over the records on disk for the Problem."""
filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split))
example_spec = problem.example_reading_spec()[0]
... | [
"def",
"tfrecord_iterator_for_problem",
"(",
"problem",
",",
"data_dir",
",",
"dataset_split",
"=",
"tf",
".",
"estimator",
".",
"ModeKeys",
".",
"TRAIN",
")",
":",
"filenames",
"=",
"tf",
".",
"gfile",
".",
"Glob",
"(",
"problem",
".",
"filepattern",
"(",
... | Iterate over the records on disk for the Problem. | [
"Iterate",
"over",
"the",
"records",
"on",
"disk",
"for",
"the",
"Problem",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L898-L903 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | tfrecord_iterator | def tfrecord_iterator(filenames, gzipped=False, example_spec=None):
"""Yields records from TFRecord files.
Args:
filenames: list<str>, list of TFRecord filenames to read from.
gzipped: bool, whether the TFRecord files are gzip-encoded.
example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFe... | python | def tfrecord_iterator(filenames, gzipped=False, example_spec=None):
"""Yields records from TFRecord files.
Args:
filenames: list<str>, list of TFRecord filenames to read from.
gzipped: bool, whether the TFRecord files are gzip-encoded.
example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFe... | [
"def",
"tfrecord_iterator",
"(",
"filenames",
",",
"gzipped",
"=",
"False",
",",
"example_spec",
"=",
"None",
")",
":",
"with",
"tf",
".",
"Graph",
"(",
")",
".",
"as_default",
"(",
")",
":",
"dataset",
"=",
"tf",
".",
"data",
".",
"Dataset",
".",
"f... | Yields records from TFRecord files.
Args:
filenames: list<str>, list of TFRecord filenames to read from.
gzipped: bool, whether the TFRecord files are gzip-encoded.
example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFeature>,
if provided, will parse each record as a tensorflow.Example... | [
"Yields",
"records",
"from",
"TFRecord",
"files",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L906-L943 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/generator_utils.py | random_deinterleave | def random_deinterleave(text, separator_symbol="X"):
"""Create a fill-in-the-blanks training example from text.
Split on spaces, then cut into segments at random points. Alternate segments
are assigned to the two output strings. separator_symbol separates segments
within each of the outputs.
example:
t... | python | def random_deinterleave(text, separator_symbol="X"):
"""Create a fill-in-the-blanks training example from text.
Split on spaces, then cut into segments at random points. Alternate segments
are assigned to the two output strings. separator_symbol separates segments
within each of the outputs.
example:
t... | [
"def",
"random_deinterleave",
"(",
"text",
",",
"separator_symbol",
"=",
"\"X\"",
")",
":",
"words",
"=",
"text",
".",
"strip",
"(",
")",
".",
"split",
"(",
"\" \"",
")",
"n",
"=",
"len",
"(",
"words",
")",
"if",
"n",
"<=",
"1",
":",
"return",
"tex... | Create a fill-in-the-blanks training example from text.
Split on spaces, then cut into segments at random points. Alternate segments
are assigned to the two output strings. separator_symbol separates segments
within each of the outputs.
example:
text="The quick brown fox jumps over the lazy dog."
ret... | [
"Create",
"a",
"fill",
"-",
"in",
"-",
"the",
"-",
"blanks",
"training",
"example",
"from",
"text",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L946-L981 | train |
tensorflow/tensor2tensor | tensor2tensor/models/neural_gpu.py | neural_gpu_body | def neural_gpu_body(inputs, hparams, name=None):
"""The core Neural GPU."""
with tf.variable_scope(name, "neural_gpu"):
def step(state, inp): # pylint: disable=missing-docstring
x = tf.nn.dropout(state, 1.0 - hparams.dropout)
for layer in range(hparams.num_hidden_layers):
x = common_layers... | python | def neural_gpu_body(inputs, hparams, name=None):
"""The core Neural GPU."""
with tf.variable_scope(name, "neural_gpu"):
def step(state, inp): # pylint: disable=missing-docstring
x = tf.nn.dropout(state, 1.0 - hparams.dropout)
for layer in range(hparams.num_hidden_layers):
x = common_layers... | [
"def",
"neural_gpu_body",
"(",
"inputs",
",",
"hparams",
",",
"name",
"=",
"None",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"name",
",",
"\"neural_gpu\"",
")",
":",
"def",
"step",
"(",
"state",
",",
"inp",
")",
":",
"# pylint: disable=missing-do... | The core Neural GPU. | [
"The",
"core",
"Neural",
"GPU",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/neural_gpu.py#L31-L52 | train |
tensorflow/tensor2tensor | tensor2tensor/models/neural_gpu.py | diagonal_neural_gpu | def diagonal_neural_gpu(inputs, hparams, name=None):
"""Improved Neural GPU as in https://arxiv.org/abs/1702.08727."""
with tf.variable_scope(name, "diagonal_neural_gpu"):
def step(state_tup, inp):
"""Single step of the improved Neural GPU."""
state, _ = state_tup
x = state
for layer in... | python | def diagonal_neural_gpu(inputs, hparams, name=None):
"""Improved Neural GPU as in https://arxiv.org/abs/1702.08727."""
with tf.variable_scope(name, "diagonal_neural_gpu"):
def step(state_tup, inp):
"""Single step of the improved Neural GPU."""
state, _ = state_tup
x = state
for layer in... | [
"def",
"diagonal_neural_gpu",
"(",
"inputs",
",",
"hparams",
",",
"name",
"=",
"None",
")",
":",
"with",
"tf",
".",
"variable_scope",
"(",
"name",
",",
"\"diagonal_neural_gpu\"",
")",
":",
"def",
"step",
"(",
"state_tup",
",",
"inp",
")",
":",
"\"\"\"Singl... | Improved Neural GPU as in https://arxiv.org/abs/1702.08727. | [
"Improved",
"Neural",
"GPU",
"as",
"in",
"https",
":",
"//",
"arxiv",
".",
"org",
"/",
"abs",
"/",
"1702",
".",
"08727",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/neural_gpu.py#L62-L87 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | _reorder_shape | def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name
"""Helper to determine the shape of reorder output."""
if output is None:
return input_shape
return base.nested_map(output, lambda i: input_shape[i]) | python | def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name
"""Helper to determine the shape of reorder output."""
if output is None:
return input_shape
return base.nested_map(output, lambda i: input_shape[i]) | [
"def",
"_reorder_shape",
"(",
"input_shape",
",",
"output",
"=",
"None",
")",
":",
"# pylint: disable=invalid-name",
"if",
"output",
"is",
"None",
":",
"return",
"input_shape",
"return",
"base",
".",
"nested_map",
"(",
"output",
",",
"lambda",
"i",
":",
"input... | Helper to determine the shape of reorder output. | [
"Helper",
"to",
"determine",
"the",
"shape",
"of",
"reorder",
"output",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L77-L81 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | Reorder | def Reorder(x, params, output=None, **kwargs):
"""Reorder a tuple into another tuple.
For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y).
The output argument specifies how to re-order, using integers that refer
to indices in the input tuple. For example, if
input = (x, y, z)
then
... | python | def Reorder(x, params, output=None, **kwargs):
"""Reorder a tuple into another tuple.
For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y).
The output argument specifies how to re-order, using integers that refer
to indices in the input tuple. For example, if
input = (x, y, z)
then
... | [
"def",
"Reorder",
"(",
"x",
",",
"params",
",",
"output",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"del",
"params",
",",
"kwargs",
"if",
"output",
"is",
"None",
":",
"return",
"x",
"return",
"base",
".",
"nested_map",
"(",
"output",
",",
"la... | Reorder a tuple into another tuple.
For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y).
The output argument specifies how to re-order, using integers that refer
to indices in the input tuple. For example, if
input = (x, y, z)
then
Reorder(input, output=(1, 0, 2)) = (y, x, z)
... | [
"Reorder",
"a",
"tuple",
"into",
"another",
"tuple",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L85-L115 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | _nested_op | def _nested_op(inputs, op): # pylint: disable=invalid-name
"""Helper: sum a list of arrays or nested arrays."""
# First the simple non-nested case.
if not isinstance(inputs[0], (list, tuple)):
return op(inputs)
# In the nested case, sum on each axis separately.
result_list = []
for i in range(len(input... | python | def _nested_op(inputs, op): # pylint: disable=invalid-name
"""Helper: sum a list of arrays or nested arrays."""
# First the simple non-nested case.
if not isinstance(inputs[0], (list, tuple)):
return op(inputs)
# In the nested case, sum on each axis separately.
result_list = []
for i in range(len(input... | [
"def",
"_nested_op",
"(",
"inputs",
",",
"op",
")",
":",
"# pylint: disable=invalid-name",
"# First the simple non-nested case.",
"if",
"not",
"isinstance",
"(",
"inputs",
"[",
"0",
"]",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"return",
"op",
"(",
"in... | Helper: sum a list of arrays or nested arrays. | [
"Helper",
":",
"sum",
"a",
"list",
"of",
"arrays",
"or",
"nested",
"arrays",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L134-L145 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | GateBranches | def GateBranches(x, **unused_kwargs):
"""Implements a gating function on a (memory, gate, candidate) tuple.
Final update is memory * gate + (1-gate) * candidate
This gating equation may also be referred to as Highway Network.
Highway Networks: https://arxiv.org/abs/1505.00387
Args:
x: A tuple of (memor... | python | def GateBranches(x, **unused_kwargs):
"""Implements a gating function on a (memory, gate, candidate) tuple.
Final update is memory * gate + (1-gate) * candidate
This gating equation may also be referred to as Highway Network.
Highway Networks: https://arxiv.org/abs/1505.00387
Args:
x: A tuple of (memor... | [
"def",
"GateBranches",
"(",
"x",
",",
"*",
"*",
"unused_kwargs",
")",
":",
"assert",
"len",
"(",
"x",
")",
"==",
"3",
",",
"x",
"state",
",",
"gate",
",",
"candidate",
"=",
"x",
"return",
"gate",
"*",
"state",
"+",
"(",
"1.0",
"-",
"gate",
")",
... | Implements a gating function on a (memory, gate, candidate) tuple.
Final update is memory * gate + (1-gate) * candidate
This gating equation may also be referred to as Highway Network.
Highway Networks: https://arxiv.org/abs/1505.00387
Args:
x: A tuple of (memory, gate, candidate)
Returns:
The res... | [
"Implements",
"a",
"gating",
"function",
"on",
"a",
"(",
"memory",
"gate",
"candidate",
")",
"tuple",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L170-L186 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | _concatenate_shape | def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name
"""Helper to determine the shape of Concatenate output."""
ax = axis % len(input_shape[0])
concat_size = sum(shape[ax] for shape in input_shape)
out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:]
return out_sha... | python | def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name
"""Helper to determine the shape of Concatenate output."""
ax = axis % len(input_shape[0])
concat_size = sum(shape[ax] for shape in input_shape)
out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:]
return out_sha... | [
"def",
"_concatenate_shape",
"(",
"input_shape",
",",
"axis",
"=",
"-",
"1",
")",
":",
"# pylint: disable=invalid-name",
"ax",
"=",
"axis",
"%",
"len",
"(",
"input_shape",
"[",
"0",
"]",
")",
"concat_size",
"=",
"sum",
"(",
"shape",
"[",
"ax",
"]",
"for"... | Helper to determine the shape of Concatenate output. | [
"Helper",
"to",
"determine",
"the",
"shape",
"of",
"Concatenate",
"output",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L189-L194 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/layers/combinators.py | Residual | def Residual(*layers, **kwargs):
"""Constructs a residual version of layers, summing input to layers output."""
shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter
if len(layers) > 1:
return Serial(
Branch(), # pylint: disable=no-value-for-parameter
Paralle... | python | def Residual(*layers, **kwargs):
"""Constructs a residual version of layers, summing input to layers output."""
shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter
if len(layers) > 1:
return Serial(
Branch(), # pylint: disable=no-value-for-parameter
Paralle... | [
"def",
"Residual",
"(",
"*",
"layers",
",",
"*",
"*",
"kwargs",
")",
":",
"shortcut",
"=",
"kwargs",
".",
"get",
"(",
"'shortcut'",
",",
"Identity",
"(",
")",
")",
"# pylint: disable=no-value-for-parameter",
"if",
"len",
"(",
"layers",
")",
">",
"1",
":"... | Constructs a residual version of layers, summing input to layers output. | [
"Constructs",
"a",
"residual",
"version",
"of",
"layers",
"summing",
"input",
"to",
"layers",
"output",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L240-L256 | train |
tensorflow/tensor2tensor | tensor2tensor/rl/policy_learner.py | PolicyLearner.train | def train(
self,
env_fn,
hparams,
simulated,
save_continuously,
epoch,
sampling_temp=1.0,
num_env_steps=None,
env_step_multiplier=1,
eval_env_fn=None,
report_fn=None
):
"""Train."""
raise NotImplementedError() | python | def train(
self,
env_fn,
hparams,
simulated,
save_continuously,
epoch,
sampling_temp=1.0,
num_env_steps=None,
env_step_multiplier=1,
eval_env_fn=None,
report_fn=None
):
"""Train."""
raise NotImplementedError() | [
"def",
"train",
"(",
"self",
",",
"env_fn",
",",
"hparams",
",",
"simulated",
",",
"save_continuously",
",",
"epoch",
",",
"sampling_temp",
"=",
"1.0",
",",
"num_env_steps",
"=",
"None",
",",
"env_step_multiplier",
"=",
"1",
",",
"eval_env_fn",
"=",
"None",
... | Train. | [
"Train",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/rl/policy_learner.py#L34-L48 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | update_hparams_for_universal_transformer | def update_hparams_for_universal_transformer(hparams):
"""Adds default hparams for all of the variants of the Universal Transformer.
Args:
hparams: default hparams (usually one of the standard hparams from
transformer model (like "transformer_base")
Returns:
hparams with default values for Univers... | python | def update_hparams_for_universal_transformer(hparams):
"""Adds default hparams for all of the variants of the Universal Transformer.
Args:
hparams: default hparams (usually one of the standard hparams from
transformer model (like "transformer_base")
Returns:
hparams with default values for Univers... | [
"def",
"update_hparams_for_universal_transformer",
"(",
"hparams",
")",
":",
"hparams",
".",
"daisy_chain_variables",
"=",
"False",
"# Breaks multi-gpu in while loops.",
"# If not None, mixes vanilla transformer with Universal Transformer.",
"# Options: None, \"before_ut\", and \"after_ut\... | Adds default hparams for all of the variants of the Universal Transformer.
Args:
hparams: default hparams (usually one of the standard hparams from
transformer model (like "transformer_base")
Returns:
hparams with default values for Universal Transformers hyper-parameters | [
"Adds",
"default",
"hparams",
"for",
"all",
"of",
"the",
"variants",
"of",
"the",
"Universal",
"Transformer",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L352-L436 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | universal_transformer_base | def universal_transformer_base():
"""Base parameters for Universal Transformer."""
hparams = transformer.transformer_base()
# To have a similar capacity to the transformer_base with 6 layers,
# we need to increase the size of the UT's layer
# since, in fact, UT has a single layer repeating multiple times.
h... | python | def universal_transformer_base():
"""Base parameters for Universal Transformer."""
hparams = transformer.transformer_base()
# To have a similar capacity to the transformer_base with 6 layers,
# we need to increase the size of the UT's layer
# since, in fact, UT has a single layer repeating multiple times.
h... | [
"def",
"universal_transformer_base",
"(",
")",
":",
"hparams",
"=",
"transformer",
".",
"transformer_base",
"(",
")",
"# To have a similar capacity to the transformer_base with 6 layers,",
"# we need to increase the size of the UT's layer",
"# since, in fact, UT has a single layer repeat... | Base parameters for Universal Transformer. | [
"Base",
"parameters",
"for",
"Universal",
"Transformer",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L440-L451 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | adaptive_universal_transformer_multilayer_tpu | def adaptive_universal_transformer_multilayer_tpu():
"""Multi-layer config for adaptive Transformer on TPU."""
hparams = adaptive_universal_transformer_base_tpu()
hparams.num_inrecurrence_layers = 2
hparams.mix_with_transformer = "before_ut,after_ut"
hparams.num_mixedin_layers = 1
hparams.transformer_ffn_ty... | python | def adaptive_universal_transformer_multilayer_tpu():
"""Multi-layer config for adaptive Transformer on TPU."""
hparams = adaptive_universal_transformer_base_tpu()
hparams.num_inrecurrence_layers = 2
hparams.mix_with_transformer = "before_ut,after_ut"
hparams.num_mixedin_layers = 1
hparams.transformer_ffn_ty... | [
"def",
"adaptive_universal_transformer_multilayer_tpu",
"(",
")",
":",
"hparams",
"=",
"adaptive_universal_transformer_base_tpu",
"(",
")",
"hparams",
".",
"num_inrecurrence_layers",
"=",
"2",
"hparams",
".",
"mix_with_transformer",
"=",
"\"before_ut,after_ut\"",
"hparams",
... | Multi-layer config for adaptive Transformer on TPU. | [
"Multi",
"-",
"layer",
"config",
"for",
"adaptive",
"Transformer",
"on",
"TPU",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L543-L555 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | adaptive_universal_transformer_multilayer_hard | def adaptive_universal_transformer_multilayer_hard():
"""Multi-layer config for adaptive Transformer with hard attention."""
hparams = adaptive_universal_transformer_multilayer_tpu()
hparams.batch_size = 256
hparams.hard_attention_k = 8
hparams.add_step_timing_signal = True
# hparams.add_sru = True # This ... | python | def adaptive_universal_transformer_multilayer_hard():
"""Multi-layer config for adaptive Transformer with hard attention."""
hparams = adaptive_universal_transformer_multilayer_tpu()
hparams.batch_size = 256
hparams.hard_attention_k = 8
hparams.add_step_timing_signal = True
# hparams.add_sru = True # This ... | [
"def",
"adaptive_universal_transformer_multilayer_hard",
"(",
")",
":",
"hparams",
"=",
"adaptive_universal_transformer_multilayer_tpu",
"(",
")",
"hparams",
".",
"batch_size",
"=",
"256",
"hparams",
".",
"hard_attention_k",
"=",
"8",
"hparams",
".",
"add_step_timing_sign... | Multi-layer config for adaptive Transformer with hard attention. | [
"Multi",
"-",
"layer",
"config",
"for",
"adaptive",
"Transformer",
"with",
"hard",
"attention",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L559-L568 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | universal_transformer_base_range | def universal_transformer_base_range(rhp):
"""Range of hyperparameters."""
# After starting from base, set intervals for some parameters.
rhp.set_discrete("num_rec_steps", [6, 8, 10])
rhp.set_discrete("hidden_size", [1024, 2048, 4096])
rhp.set_discrete("filter_size", [2048, 4096, 8192])
rhp.set_discrete("nu... | python | def universal_transformer_base_range(rhp):
"""Range of hyperparameters."""
# After starting from base, set intervals for some parameters.
rhp.set_discrete("num_rec_steps", [6, 8, 10])
rhp.set_discrete("hidden_size", [1024, 2048, 4096])
rhp.set_discrete("filter_size", [2048, 4096, 8192])
rhp.set_discrete("nu... | [
"def",
"universal_transformer_base_range",
"(",
"rhp",
")",
":",
"# After starting from base, set intervals for some parameters.",
"rhp",
".",
"set_discrete",
"(",
"\"num_rec_steps\"",
",",
"[",
"6",
",",
"8",
",",
"10",
"]",
")",
"rhp",
".",
"set_discrete",
"(",
"\... | Range of hyperparameters. | [
"Range",
"of",
"hyperparameters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L788-L797 | train |
tensorflow/tensor2tensor | tensor2tensor/models/research/universal_transformer.py | adaptive_universal_transformer_base_range | def adaptive_universal_transformer_base_range(rhp):
"""Range of hyperparameters."""
# After starting from base, set intervals for some parameters.
rhp.set_discrete("act_max_steps", [8, 16, 32])
rhp.set_float("act_loss_weight", 0.0, 0.5)
rhp.set_discrete("hidden_size", [1024, 2048, 4096])
rhp.set_discrete("f... | python | def adaptive_universal_transformer_base_range(rhp):
"""Range of hyperparameters."""
# After starting from base, set intervals for some parameters.
rhp.set_discrete("act_max_steps", [8, 16, 32])
rhp.set_float("act_loss_weight", 0.0, 0.5)
rhp.set_discrete("hidden_size", [1024, 2048, 4096])
rhp.set_discrete("f... | [
"def",
"adaptive_universal_transformer_base_range",
"(",
"rhp",
")",
":",
"# After starting from base, set intervals for some parameters.",
"rhp",
".",
"set_discrete",
"(",
"\"act_max_steps\"",
",",
"[",
"8",
",",
"16",
",",
"32",
"]",
")",
"rhp",
".",
"set_float",
"(... | Range of hyperparameters. | [
"Range",
"of",
"hyperparameters",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L801-L811 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/neural_gpu.py | DiagonalGate | def DiagonalGate(x, params, **kwargs):
"""Split channels in 3 parts. Shifts 1st and 3rd sections to left/right."""
del params
del kwargs
# x : [batch, 1, length, depth]
x = np.pad(
x, [(0, 0), (0, 0), (1, 1), (0, 0)], mode='constant', constant_values=0.0)
depth = x.shape[-1] // 3
assert 3 * depth ==... | python | def DiagonalGate(x, params, **kwargs):
"""Split channels in 3 parts. Shifts 1st and 3rd sections to left/right."""
del params
del kwargs
# x : [batch, 1, length, depth]
x = np.pad(
x, [(0, 0), (0, 0), (1, 1), (0, 0)], mode='constant', constant_values=0.0)
depth = x.shape[-1] // 3
assert 3 * depth ==... | [
"def",
"DiagonalGate",
"(",
"x",
",",
"params",
",",
"*",
"*",
"kwargs",
")",
":",
"del",
"params",
"del",
"kwargs",
"# x : [batch, 1, length, depth]",
"x",
"=",
"np",
".",
"pad",
"(",
"x",
",",
"[",
"(",
"0",
",",
"0",
")",
",",
"(",
"0",
",",
"... | Split channels in 3 parts. Shifts 1st and 3rd sections to left/right. | [
"Split",
"channels",
"in",
"3",
"parts",
".",
"Shifts",
"1st",
"and",
"3rd",
"sections",
"to",
"left",
"/",
"right",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L33-L47 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/neural_gpu.py | ConvDiagonalGRU | def ConvDiagonalGRU(units, kernel_size=(3, 3)):
"""Build convolutional GRU with diagonal gating as in ImprovedNGPU."""
def BuildConv():
return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME')
return layers.GeneralGRUCell(
candidate_transform=BuildConv,
memory_transform=Diagona... | python | def ConvDiagonalGRU(units, kernel_size=(3, 3)):
"""Build convolutional GRU with diagonal gating as in ImprovedNGPU."""
def BuildConv():
return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME')
return layers.GeneralGRUCell(
candidate_transform=BuildConv,
memory_transform=Diagona... | [
"def",
"ConvDiagonalGRU",
"(",
"units",
",",
"kernel_size",
"=",
"(",
"3",
",",
"3",
")",
")",
":",
"def",
"BuildConv",
"(",
")",
":",
"return",
"layers",
".",
"Conv",
"(",
"filters",
"=",
"units",
",",
"kernel_size",
"=",
"kernel_size",
",",
"padding"... | Build convolutional GRU with diagonal gating as in ImprovedNGPU. | [
"Build",
"convolutional",
"GRU",
"with",
"diagonal",
"gating",
"as",
"in",
"ImprovedNGPU",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L50-L60 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/models/neural_gpu.py | NeuralGPU | def NeuralGPU(feature_depth=96, steps=16, vocab_size=2):
"""Implementation of Neural GPU: https://arxiv.org/abs/1702.08727.
Args:
feature_depth: Number of memory channels
steps: Number of times depthwise recurrence steps.
vocab_size: Vocabulary size.
Returns:
A NeuralGPU Stax model.
"""
xs =... | python | def NeuralGPU(feature_depth=96, steps=16, vocab_size=2):
"""Implementation of Neural GPU: https://arxiv.org/abs/1702.08727.
Args:
feature_depth: Number of memory channels
steps: Number of times depthwise recurrence steps.
vocab_size: Vocabulary size.
Returns:
A NeuralGPU Stax model.
"""
xs =... | [
"def",
"NeuralGPU",
"(",
"feature_depth",
"=",
"96",
",",
"steps",
"=",
"16",
",",
"vocab_size",
"=",
"2",
")",
":",
"xs",
"=",
"[",
"]",
"xs",
".",
"append",
"(",
"layers",
".",
"Embedding",
"(",
"feature_depth",
"=",
"feature_depth",
",",
"vocab_size... | Implementation of Neural GPU: https://arxiv.org/abs/1702.08727.
Args:
feature_depth: Number of memory channels
steps: Number of times depthwise recurrence steps.
vocab_size: Vocabulary size.
Returns:
A NeuralGPU Stax model. | [
"Implementation",
"of",
"Neural",
"GPU",
":",
"https",
":",
"//",
"arxiv",
".",
"org",
"/",
"abs",
"/",
"1702",
".",
"08727",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L63-L82 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | strip_ids | def strip_ids(ids, ids_to_strip):
"""Strip ids_to_strip from the end ids."""
ids = list(ids)
while ids and ids[-1] in ids_to_strip:
ids.pop()
return ids | python | def strip_ids(ids, ids_to_strip):
"""Strip ids_to_strip from the end ids."""
ids = list(ids)
while ids and ids[-1] in ids_to_strip:
ids.pop()
return ids | [
"def",
"strip_ids",
"(",
"ids",
",",
"ids_to_strip",
")",
":",
"ids",
"=",
"list",
"(",
"ids",
")",
"while",
"ids",
"and",
"ids",
"[",
"-",
"1",
"]",
"in",
"ids_to_strip",
":",
"ids",
".",
"pop",
"(",
")",
"return",
"ids"
] | Strip ids_to_strip from the end ids. | [
"Strip",
"ids_to_strip",
"from",
"the",
"end",
"ids",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L99-L104 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | _escape_token | def _escape_token(token, alphabet):
"""Escape away underscores and OOV characters and append '_'.
This allows the token to be expressed as the concatenation of a list
of subtokens from the vocabulary. The underscore acts as a sentinel
which allows us to invertibly concatenate multiple such lists.
Args:
... | python | def _escape_token(token, alphabet):
"""Escape away underscores and OOV characters and append '_'.
This allows the token to be expressed as the concatenation of a list
of subtokens from the vocabulary. The underscore acts as a sentinel
which allows us to invertibly concatenate multiple such lists.
Args:
... | [
"def",
"_escape_token",
"(",
"token",
",",
"alphabet",
")",
":",
"if",
"not",
"isinstance",
"(",
"token",
",",
"six",
".",
"text_type",
")",
":",
"raise",
"ValueError",
"(",
"\"Expected string type for token, got %s\"",
"%",
"type",
"(",
"token",
")",
")",
"... | Escape away underscores and OOV characters and append '_'.
This allows the token to be expressed as the concatenation of a list
of subtokens from the vocabulary. The underscore acts as a sentinel
which allows us to invertibly concatenate multiple such lists.
Args:
token: A unicode string to be escaped.
... | [
"Escape",
"away",
"underscores",
"and",
"OOV",
"characters",
"and",
"append",
"_",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L400-L422 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TextEncoder.encode | def encode(self, s):
"""Transform a human-readable string into a sequence of int ids.
The ids should be in the range [num_reserved_ids, vocab_size). Ids [0,
num_reserved_ids) are reserved.
EOS is not appended.
Args:
s: human-readable string to be converted.
Returns:
ids: list of ... | python | def encode(self, s):
"""Transform a human-readable string into a sequence of int ids.
The ids should be in the range [num_reserved_ids, vocab_size). Ids [0,
num_reserved_ids) are reserved.
EOS is not appended.
Args:
s: human-readable string to be converted.
Returns:
ids: list of ... | [
"def",
"encode",
"(",
"self",
",",
"s",
")",
":",
"return",
"[",
"int",
"(",
"w",
")",
"+",
"self",
".",
"_num_reserved_ids",
"for",
"w",
"in",
"s",
".",
"split",
"(",
")",
"]"
] | Transform a human-readable string into a sequence of int ids.
The ids should be in the range [num_reserved_ids, vocab_size). Ids [0,
num_reserved_ids) are reserved.
EOS is not appended.
Args:
s: human-readable string to be converted.
Returns:
ids: list of integers | [
"Transform",
"a",
"human",
"-",
"readable",
"string",
"into",
"a",
"sequence",
"of",
"int",
"ids",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L117-L131 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TextEncoder.decode | def decode(self, ids, strip_extraneous=False):
"""Transform a sequence of int ids into a human-readable string.
EOS is not expected in ids.
Args:
ids: list of integers to be converted.
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
s: ... | python | def decode(self, ids, strip_extraneous=False):
"""Transform a sequence of int ids into a human-readable string.
EOS is not expected in ids.
Args:
ids: list of integers to be converted.
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
s: ... | [
"def",
"decode",
"(",
"self",
",",
"ids",
",",
"strip_extraneous",
"=",
"False",
")",
":",
"if",
"strip_extraneous",
":",
"ids",
"=",
"strip_ids",
"(",
"ids",
",",
"list",
"(",
"range",
"(",
"self",
".",
"_num_reserved_ids",
"or",
"0",
")",
")",
")",
... | Transform a sequence of int ids into a human-readable string.
EOS is not expected in ids.
Args:
ids: list of integers to be converted.
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
s: human-readable string. | [
"Transform",
"a",
"sequence",
"of",
"int",
"ids",
"into",
"a",
"human",
"-",
"readable",
"string",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L133-L148 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TextEncoder.decode_list | def decode_list(self, ids):
"""Transform a sequence of int ids into a their string versions.
This method supports transforming individual input/output ids to their
string versions so that sequence to/from text conversions can be visualized
in a human readable format.
Args:
ids: list of integ... | python | def decode_list(self, ids):
"""Transform a sequence of int ids into a their string versions.
This method supports transforming individual input/output ids to their
string versions so that sequence to/from text conversions can be visualized
in a human readable format.
Args:
ids: list of integ... | [
"def",
"decode_list",
"(",
"self",
",",
"ids",
")",
":",
"decoded_ids",
"=",
"[",
"]",
"for",
"id_",
"in",
"ids",
":",
"if",
"0",
"<=",
"id_",
"<",
"self",
".",
"_num_reserved_ids",
":",
"decoded_ids",
".",
"append",
"(",
"RESERVED_TOKENS",
"[",
"int",... | Transform a sequence of int ids into a their string versions.
This method supports transforming individual input/output ids to their
string versions so that sequence to/from text conversions can be visualized
in a human readable format.
Args:
ids: list of integers to be converted.
Returns:
... | [
"Transform",
"a",
"sequence",
"of",
"int",
"ids",
"into",
"a",
"their",
"string",
"versions",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L150-L169 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TokenTextEncoder.encode | def encode(self, s):
"""Converts a space-separated string of tokens to a list of ids."""
sentence = s
tokens = sentence.strip().split()
if self._replace_oov is not None:
tokens = [t if t in self._token_to_id else self._replace_oov
for t in tokens]
ret = [self._token_to_id[tok] ... | python | def encode(self, s):
"""Converts a space-separated string of tokens to a list of ids."""
sentence = s
tokens = sentence.strip().split()
if self._replace_oov is not None:
tokens = [t if t in self._token_to_id else self._replace_oov
for t in tokens]
ret = [self._token_to_id[tok] ... | [
"def",
"encode",
"(",
"self",
",",
"s",
")",
":",
"sentence",
"=",
"s",
"tokens",
"=",
"sentence",
".",
"strip",
"(",
")",
".",
"split",
"(",
")",
"if",
"self",
".",
"_replace_oov",
"is",
"not",
"None",
":",
"tokens",
"=",
"[",
"t",
"if",
"t",
... | Converts a space-separated string of tokens to a list of ids. | [
"Converts",
"a",
"space",
"-",
"separated",
"string",
"of",
"tokens",
"to",
"a",
"list",
"of",
"ids",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L314-L322 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TokenTextEncoder._init_vocab_from_file | def _init_vocab_from_file(self, filename):
"""Load vocab from a file.
Args:
filename: The file to load vocabulary from.
"""
with tf.gfile.Open(filename) as f:
tokens = [token.strip() for token in f.readlines()]
def token_gen():
for token in tokens:
yield token
self._... | python | def _init_vocab_from_file(self, filename):
"""Load vocab from a file.
Args:
filename: The file to load vocabulary from.
"""
with tf.gfile.Open(filename) as f:
tokens = [token.strip() for token in f.readlines()]
def token_gen():
for token in tokens:
yield token
self._... | [
"def",
"_init_vocab_from_file",
"(",
"self",
",",
"filename",
")",
":",
"with",
"tf",
".",
"gfile",
".",
"Open",
"(",
"filename",
")",
"as",
"f",
":",
"tokens",
"=",
"[",
"token",
".",
"strip",
"(",
")",
"for",
"token",
"in",
"f",
".",
"readlines",
... | Load vocab from a file.
Args:
filename: The file to load vocabulary from. | [
"Load",
"vocab",
"from",
"a",
"file",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L338-L351 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TokenTextEncoder._init_vocab_from_list | def _init_vocab_from_list(self, vocab_list):
"""Initialize tokens from a list of tokens.
It is ok if reserved tokens appear in the vocab list. They will be
removed. The set of tokens in vocab_list should be unique.
Args:
vocab_list: A list of tokens.
"""
def token_gen():
for token ... | python | def _init_vocab_from_list(self, vocab_list):
"""Initialize tokens from a list of tokens.
It is ok if reserved tokens appear in the vocab list. They will be
removed. The set of tokens in vocab_list should be unique.
Args:
vocab_list: A list of tokens.
"""
def token_gen():
for token ... | [
"def",
"_init_vocab_from_list",
"(",
"self",
",",
"vocab_list",
")",
":",
"def",
"token_gen",
"(",
")",
":",
"for",
"token",
"in",
"vocab_list",
":",
"if",
"token",
"not",
"in",
"RESERVED_TOKENS",
":",
"yield",
"token",
"self",
".",
"_init_vocab",
"(",
"to... | Initialize tokens from a list of tokens.
It is ok if reserved tokens appear in the vocab list. They will be
removed. The set of tokens in vocab_list should be unique.
Args:
vocab_list: A list of tokens. | [
"Initialize",
"tokens",
"from",
"a",
"list",
"of",
"tokens",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L353-L367 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TokenTextEncoder._init_vocab | def _init_vocab(self, token_generator, add_reserved_tokens=True):
"""Initialize vocabulary with tokens from token_generator."""
self._id_to_token = {}
non_reserved_start_index = 0
if add_reserved_tokens:
self._id_to_token.update(enumerate(RESERVED_TOKENS))
non_reserved_start_index = len(RE... | python | def _init_vocab(self, token_generator, add_reserved_tokens=True):
"""Initialize vocabulary with tokens from token_generator."""
self._id_to_token = {}
non_reserved_start_index = 0
if add_reserved_tokens:
self._id_to_token.update(enumerate(RESERVED_TOKENS))
non_reserved_start_index = len(RE... | [
"def",
"_init_vocab",
"(",
"self",
",",
"token_generator",
",",
"add_reserved_tokens",
"=",
"True",
")",
":",
"self",
".",
"_id_to_token",
"=",
"{",
"}",
"non_reserved_start_index",
"=",
"0",
"if",
"add_reserved_tokens",
":",
"self",
".",
"_id_to_token",
".",
... | Initialize vocabulary with tokens from token_generator. | [
"Initialize",
"vocabulary",
"with",
"tokens",
"from",
"token_generator",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L369-L384 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | TokenTextEncoder.store_to_file | def store_to_file(self, filename):
"""Write vocab file to disk.
Vocab files have one token per line. The file ends in a newline. Reserved
tokens are written to the vocab file as well.
Args:
filename: Full path of the file to store the vocab to.
"""
with tf.gfile.Open(filename, "w") as f:... | python | def store_to_file(self, filename):
"""Write vocab file to disk.
Vocab files have one token per line. The file ends in a newline. Reserved
tokens are written to the vocab file as well.
Args:
filename: Full path of the file to store the vocab to.
"""
with tf.gfile.Open(filename, "w") as f:... | [
"def",
"store_to_file",
"(",
"self",
",",
"filename",
")",
":",
"with",
"tf",
".",
"gfile",
".",
"Open",
"(",
"filename",
",",
"\"w\"",
")",
"as",
"f",
":",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"self",
".",
"_id_to_token",
")",
")",
":",
"... | Write vocab file to disk.
Vocab files have one token per line. The file ends in a newline. Reserved
tokens are written to the vocab file as well.
Args:
filename: Full path of the file to store the vocab to. | [
"Write",
"vocab",
"file",
"to",
"disk",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L386-L397 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder.decode | def decode(self, ids, strip_extraneous=False):
"""Converts a sequence of subtoken ids to a native string.
Args:
ids: a list of integers in the range [0, vocab_size)
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
a native string
"""
... | python | def decode(self, ids, strip_extraneous=False):
"""Converts a sequence of subtoken ids to a native string.
Args:
ids: a list of integers in the range [0, vocab_size)
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
a native string
"""
... | [
"def",
"decode",
"(",
"self",
",",
"ids",
",",
"strip_extraneous",
"=",
"False",
")",
":",
"if",
"strip_extraneous",
":",
"ids",
"=",
"strip_ids",
"(",
"ids",
",",
"list",
"(",
"range",
"(",
"self",
".",
"_num_reserved_ids",
"or",
"0",
")",
")",
")",
... | Converts a sequence of subtoken ids to a native string.
Args:
ids: a list of integers in the range [0, vocab_size)
strip_extraneous: bool, whether to strip off extraneous tokens
(EOS and PAD).
Returns:
a native string | [
"Converts",
"a",
"sequence",
"of",
"subtoken",
"ids",
"to",
"a",
"native",
"string",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L522-L536 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._tokens_to_subtoken_ids | def _tokens_to_subtoken_ids(self, tokens):
"""Converts a list of tokens to a list of subtoken ids.
Args:
tokens: a list of strings.
Returns:
a list of integers in the range [0, vocab_size)
"""
ret = []
for token in tokens:
ret.extend(self._token_to_subtoken_ids(token))
ret... | python | def _tokens_to_subtoken_ids(self, tokens):
"""Converts a list of tokens to a list of subtoken ids.
Args:
tokens: a list of strings.
Returns:
a list of integers in the range [0, vocab_size)
"""
ret = []
for token in tokens:
ret.extend(self._token_to_subtoken_ids(token))
ret... | [
"def",
"_tokens_to_subtoken_ids",
"(",
"self",
",",
"tokens",
")",
":",
"ret",
"=",
"[",
"]",
"for",
"token",
"in",
"tokens",
":",
"ret",
".",
"extend",
"(",
"self",
".",
"_token_to_subtoken_ids",
"(",
"token",
")",
")",
"return",
"ret"
] | Converts a list of tokens to a list of subtoken ids.
Args:
tokens: a list of strings.
Returns:
a list of integers in the range [0, vocab_size) | [
"Converts",
"a",
"list",
"of",
"tokens",
"to",
"a",
"list",
"of",
"subtoken",
"ids",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L546-L557 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._token_to_subtoken_ids | def _token_to_subtoken_ids(self, token):
"""Converts token to a list of subtoken ids.
Args:
token: a string.
Returns:
a list of integers in the range [0, vocab_size)
"""
cache_location = hash(token) % self._cache_size
cache_key, cache_value = self._cache[cache_location]
if cache... | python | def _token_to_subtoken_ids(self, token):
"""Converts token to a list of subtoken ids.
Args:
token: a string.
Returns:
a list of integers in the range [0, vocab_size)
"""
cache_location = hash(token) % self._cache_size
cache_key, cache_value = self._cache[cache_location]
if cache... | [
"def",
"_token_to_subtoken_ids",
"(",
"self",
",",
"token",
")",
":",
"cache_location",
"=",
"hash",
"(",
"token",
")",
"%",
"self",
".",
"_cache_size",
"cache_key",
",",
"cache_value",
"=",
"self",
".",
"_cache",
"[",
"cache_location",
"]",
"if",
"cache_key... | Converts token to a list of subtoken ids.
Args:
token: a string.
Returns:
a list of integers in the range [0, vocab_size) | [
"Converts",
"token",
"to",
"a",
"list",
"of",
"subtoken",
"ids",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L559-L574 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._subtoken_ids_to_tokens | def _subtoken_ids_to_tokens(self, subtokens):
"""Converts a list of subtoken ids to a list of tokens.
Args:
subtokens: a list of integers in the range [0, vocab_size)
Returns:
a list of strings.
"""
concatenated = "".join(
[self._subtoken_id_to_subtoken_string(s) for s in subtok... | python | def _subtoken_ids_to_tokens(self, subtokens):
"""Converts a list of subtoken ids to a list of tokens.
Args:
subtokens: a list of integers in the range [0, vocab_size)
Returns:
a list of strings.
"""
concatenated = "".join(
[self._subtoken_id_to_subtoken_string(s) for s in subtok... | [
"def",
"_subtoken_ids_to_tokens",
"(",
"self",
",",
"subtokens",
")",
":",
"concatenated",
"=",
"\"\"",
".",
"join",
"(",
"[",
"self",
".",
"_subtoken_id_to_subtoken_string",
"(",
"s",
")",
"for",
"s",
"in",
"subtokens",
"]",
")",
"split",
"=",
"concatenated... | Converts a list of subtoken ids to a list of tokens.
Args:
subtokens: a list of integers in the range [0, vocab_size)
Returns:
a list of strings. | [
"Converts",
"a",
"list",
"of",
"subtoken",
"ids",
"to",
"a",
"list",
"of",
"tokens",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L576-L593 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._subtoken_id_to_subtoken_string | def _subtoken_id_to_subtoken_string(self, subtoken):
"""Converts a subtoken integer ID to a subtoken string."""
if 0 <= subtoken < self.vocab_size:
return self._all_subtoken_strings[subtoken]
return u"" | python | def _subtoken_id_to_subtoken_string(self, subtoken):
"""Converts a subtoken integer ID to a subtoken string."""
if 0 <= subtoken < self.vocab_size:
return self._all_subtoken_strings[subtoken]
return u"" | [
"def",
"_subtoken_id_to_subtoken_string",
"(",
"self",
",",
"subtoken",
")",
":",
"if",
"0",
"<=",
"subtoken",
"<",
"self",
".",
"vocab_size",
":",
"return",
"self",
".",
"_all_subtoken_strings",
"[",
"subtoken",
"]",
"return",
"u\"\""
] | Converts a subtoken integer ID to a subtoken string. | [
"Converts",
"a",
"subtoken",
"integer",
"ID",
"to",
"a",
"subtoken",
"string",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L595-L599 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._escaped_token_to_subtoken_strings | def _escaped_token_to_subtoken_strings(self, escaped_token):
"""Converts an escaped token string to a list of subtoken strings.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtokens as unicode strings.
"""
# NOTE: This algorithm is greedy; it won't nece... | python | def _escaped_token_to_subtoken_strings(self, escaped_token):
"""Converts an escaped token string to a list of subtoken strings.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtokens as unicode strings.
"""
# NOTE: This algorithm is greedy; it won't nece... | [
"def",
"_escaped_token_to_subtoken_strings",
"(",
"self",
",",
"escaped_token",
")",
":",
"# NOTE: This algorithm is greedy; it won't necessarily produce the \"best\"",
"# list of subtokens.",
"ret",
"=",
"[",
"]",
"start",
"=",
"0",
"token_len",
"=",
"len",
"(",
"escaped_t... | Converts an escaped token string to a list of subtoken strings.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtokens as unicode strings. | [
"Converts",
"an",
"escaped",
"token",
"string",
"to",
"a",
"list",
"of",
"subtoken",
"strings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L601-L629 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._escaped_token_to_subtoken_ids | def _escaped_token_to_subtoken_ids(self, escaped_token):
"""Converts an escaped token string to a list of subtoken IDs.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtoken IDs as integers.
"""
return [
self._subtoken_string_to_id[subtoken]
... | python | def _escaped_token_to_subtoken_ids(self, escaped_token):
"""Converts an escaped token string to a list of subtoken IDs.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtoken IDs as integers.
"""
return [
self._subtoken_string_to_id[subtoken]
... | [
"def",
"_escaped_token_to_subtoken_ids",
"(",
"self",
",",
"escaped_token",
")",
":",
"return",
"[",
"self",
".",
"_subtoken_string_to_id",
"[",
"subtoken",
"]",
"for",
"subtoken",
"in",
"self",
".",
"_escaped_token_to_subtoken_strings",
"(",
"escaped_token",
")",
"... | Converts an escaped token string to a list of subtoken IDs.
Args:
escaped_token: An escaped token as a unicode string.
Returns:
A list of subtoken IDs as integers. | [
"Converts",
"an",
"escaped",
"token",
"string",
"to",
"a",
"list",
"of",
"subtoken",
"IDs",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L631-L642 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder.build_from_generator | def build_from_generator(cls,
generator,
target_size,
max_subtoken_length=None,
reserved_tokens=None):
"""Builds a SubwordTextEncoder from the generated text.
Args:
generator: yields text.
ta... | python | def build_from_generator(cls,
generator,
target_size,
max_subtoken_length=None,
reserved_tokens=None):
"""Builds a SubwordTextEncoder from the generated text.
Args:
generator: yields text.
ta... | [
"def",
"build_from_generator",
"(",
"cls",
",",
"generator",
",",
"target_size",
",",
"max_subtoken_length",
"=",
"None",
",",
"reserved_tokens",
"=",
"None",
")",
":",
"token_counts",
"=",
"collections",
".",
"defaultdict",
"(",
"int",
")",
"for",
"item",
"in... | Builds a SubwordTextEncoder from the generated text.
Args:
generator: yields text.
target_size: int, approximate vocabulary size to create.
max_subtoken_length: Maximum length of a subtoken. If this is not set,
then the runtime and memory use of creating the vocab is quadratic in
... | [
"Builds",
"a",
"SubwordTextEncoder",
"from",
"the",
"generated",
"text",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L645-L674 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder.build_to_target_size | def build_to_target_size(cls,
target_size,
token_counts,
min_val,
max_val,
max_subtoken_length=None,
reserved_tokens=None,
num_iter... | python | def build_to_target_size(cls,
target_size,
token_counts,
min_val,
max_val,
max_subtoken_length=None,
reserved_tokens=None,
num_iter... | [
"def",
"build_to_target_size",
"(",
"cls",
",",
"target_size",
",",
"token_counts",
",",
"min_val",
",",
"max_val",
",",
"max_subtoken_length",
"=",
"None",
",",
"reserved_tokens",
"=",
"None",
",",
"num_iterations",
"=",
"4",
")",
":",
"if",
"min_val",
">",
... | Builds a SubwordTextEncoder that has `vocab_size` near `target_size`.
Uses simple recursive binary search to find a minimum token count that most
closely matches the `target_size`.
Args:
target_size: Desired vocab_size to approximate.
token_counts: A dictionary of token counts, mapping string ... | [
"Builds",
"a",
"SubwordTextEncoder",
"that",
"has",
"vocab_size",
"near",
"target_size",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L677-L748 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder.build_from_token_counts | def build_from_token_counts(self,
token_counts,
min_count,
num_iterations=4,
reserved_tokens=None,
max_subtoken_length=None):
"""Train a SubwordTextEncoder based on a... | python | def build_from_token_counts(self,
token_counts,
min_count,
num_iterations=4,
reserved_tokens=None,
max_subtoken_length=None):
"""Train a SubwordTextEncoder based on a... | [
"def",
"build_from_token_counts",
"(",
"self",
",",
"token_counts",
",",
"min_count",
",",
"num_iterations",
"=",
"4",
",",
"reserved_tokens",
"=",
"None",
",",
"max_subtoken_length",
"=",
"None",
")",
":",
"if",
"reserved_tokens",
"is",
"None",
":",
"reserved_t... | Train a SubwordTextEncoder based on a dictionary of word counts.
Args:
token_counts: a dictionary of Unicode strings to int.
min_count: an integer - discard subtokens with lower counts.
num_iterations: an integer. how many iterations of refinement.
reserved_tokens: List of reserved tokens.... | [
"Train",
"a",
"SubwordTextEncoder",
"based",
"on",
"a",
"dictionary",
"of",
"word",
"counts",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L750-L866 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder.dump | def dump(self):
"""Debugging dump of the current subtoken vocabulary."""
subtoken_strings = [(i, s)
for s, i in six.iteritems(self._subtoken_string_to_id)]
print(u", ".join(u"{0} : '{1}'".format(i, s)
for i, s in sorted(subtoken_strings))) | python | def dump(self):
"""Debugging dump of the current subtoken vocabulary."""
subtoken_strings = [(i, s)
for s, i in six.iteritems(self._subtoken_string_to_id)]
print(u", ".join(u"{0} : '{1}'".format(i, s)
for i, s in sorted(subtoken_strings))) | [
"def",
"dump",
"(",
"self",
")",
":",
"subtoken_strings",
"=",
"[",
"(",
"i",
",",
"s",
")",
"for",
"s",
",",
"i",
"in",
"six",
".",
"iteritems",
"(",
"self",
".",
"_subtoken_string_to_id",
")",
"]",
"print",
"(",
"u\", \"",
".",
"join",
"(",
"u\"{... | Debugging dump of the current subtoken vocabulary. | [
"Debugging",
"dump",
"of",
"the",
"current",
"subtoken",
"vocabulary",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L872-L877 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._init_subtokens_from_list | def _init_subtokens_from_list(self, subtoken_strings, reserved_tokens=None):
"""Initialize token information from a list of subtoken strings.
Args:
subtoken_strings: a list of subtokens
reserved_tokens: List of reserved tokens. We must have `reserved_tokens`
as None or the empty list, or el... | python | def _init_subtokens_from_list(self, subtoken_strings, reserved_tokens=None):
"""Initialize token information from a list of subtoken strings.
Args:
subtoken_strings: a list of subtokens
reserved_tokens: List of reserved tokens. We must have `reserved_tokens`
as None or the empty list, or el... | [
"def",
"_init_subtokens_from_list",
"(",
"self",
",",
"subtoken_strings",
",",
"reserved_tokens",
"=",
"None",
")",
":",
"if",
"reserved_tokens",
"is",
"None",
":",
"reserved_tokens",
"=",
"[",
"]",
"if",
"reserved_tokens",
":",
"self",
".",
"_all_subtoken_strings... | Initialize token information from a list of subtoken strings.
Args:
subtoken_strings: a list of subtokens
reserved_tokens: List of reserved tokens. We must have `reserved_tokens`
as None or the empty list, or else the global variable `RESERVED_TOKENS`
must be a prefix of `reserved_token... | [
"Initialize",
"token",
"information",
"from",
"a",
"list",
"of",
"subtoken",
"strings",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L879-L910 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._load_from_file_object | def _load_from_file_object(self, f):
"""Load from a file object.
Args:
f: File object to load vocabulary from
"""
subtoken_strings = []
for line in f:
s = line.strip()
# Some vocab files wrap words in single quotes, but others don't
if ((s.startswith("'") and s.endswith("'")... | python | def _load_from_file_object(self, f):
"""Load from a file object.
Args:
f: File object to load vocabulary from
"""
subtoken_strings = []
for line in f:
s = line.strip()
# Some vocab files wrap words in single quotes, but others don't
if ((s.startswith("'") and s.endswith("'")... | [
"def",
"_load_from_file_object",
"(",
"self",
",",
"f",
")",
":",
"subtoken_strings",
"=",
"[",
"]",
"for",
"line",
"in",
"f",
":",
"s",
"=",
"line",
".",
"strip",
"(",
")",
"# Some vocab files wrap words in single quotes, but others don't",
"if",
"(",
"(",
"s... | Load from a file object.
Args:
f: File object to load vocabulary from | [
"Load",
"from",
"a",
"file",
"object",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L919-L934 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | SubwordTextEncoder._load_from_file | def _load_from_file(self, filename):
"""Load from a vocab file."""
if not tf.gfile.Exists(filename):
raise ValueError("File %s not found" % filename)
with tf.gfile.Open(filename) as f:
self._load_from_file_object(f) | python | def _load_from_file(self, filename):
"""Load from a vocab file."""
if not tf.gfile.Exists(filename):
raise ValueError("File %s not found" % filename)
with tf.gfile.Open(filename) as f:
self._load_from_file_object(f) | [
"def",
"_load_from_file",
"(",
"self",
",",
"filename",
")",
":",
"if",
"not",
"tf",
".",
"gfile",
".",
"Exists",
"(",
"filename",
")",
":",
"raise",
"ValueError",
"(",
"\"File %s not found\"",
"%",
"filename",
")",
"with",
"tf",
".",
"gfile",
".",
"Open... | Load from a vocab file. | [
"Load",
"from",
"a",
"vocab",
"file",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L936-L941 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | ImageEncoder.encode | def encode(self, s):
"""Transform a string with a filename into a list of RGB integers.
Args:
s: path to the file with an image.
Returns:
ids: list of integers
"""
try:
import matplotlib.image as im # pylint: disable=g-import-not-at-top
except ImportError as e:
tf.logg... | python | def encode(self, s):
"""Transform a string with a filename into a list of RGB integers.
Args:
s: path to the file with an image.
Returns:
ids: list of integers
"""
try:
import matplotlib.image as im # pylint: disable=g-import-not-at-top
except ImportError as e:
tf.logg... | [
"def",
"encode",
"(",
"self",
",",
"s",
")",
":",
"try",
":",
"import",
"matplotlib",
".",
"image",
"as",
"im",
"# pylint: disable=g-import-not-at-top",
"except",
"ImportError",
"as",
"e",
":",
"tf",
".",
"logging",
".",
"warning",
"(",
"\"Reading an image req... | Transform a string with a filename into a list of RGB integers.
Args:
s: path to the file with an image.
Returns:
ids: list of integers | [
"Transform",
"a",
"string",
"with",
"a",
"filename",
"into",
"a",
"list",
"of",
"RGB",
"integers",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L965-L980 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | ImageEncoder.decode | def decode(self, ids, strip_extraneous=False):
"""Transform a sequence of int ids into an image file.
Args:
ids: list of integers to be converted.
strip_extraneous: unused
Returns:
Path to the temporary file where the image was saved.
Raises:
ValueError: if the ids are not of ... | python | def decode(self, ids, strip_extraneous=False):
"""Transform a sequence of int ids into an image file.
Args:
ids: list of integers to be converted.
strip_extraneous: unused
Returns:
Path to the temporary file where the image was saved.
Raises:
ValueError: if the ids are not of ... | [
"def",
"decode",
"(",
"self",
",",
"ids",
",",
"strip_extraneous",
"=",
"False",
")",
":",
"del",
"strip_extraneous",
"_",
",",
"tmp_file_path",
"=",
"tempfile",
".",
"mkstemp",
"(",
"\"_decode.png\"",
")",
"if",
"self",
".",
"_height",
"is",
"None",
"or",... | Transform a sequence of int ids into an image file.
Args:
ids: list of integers to be converted.
strip_extraneous: unused
Returns:
Path to the temporary file where the image was saved.
Raises:
ValueError: if the ids are not of the appropriate size. | [
"Transform",
"a",
"sequence",
"of",
"int",
"ids",
"into",
"an",
"image",
"file",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L982-L1018 | train |
tensorflow/tensor2tensor | tensor2tensor/data_generators/text_encoder.py | RealEncoder.decode | def decode(self, ids, strip_extraneous=False):
"""Transform sequence of float values into string (float values).
Args:
ids: array of floats to be converted.
strip_extraneous: unused
Returns:
String having space separated float values.
Raises:
ValueError: if the ids are not of ... | python | def decode(self, ids, strip_extraneous=False):
"""Transform sequence of float values into string (float values).
Args:
ids: array of floats to be converted.
strip_extraneous: unused
Returns:
String having space separated float values.
Raises:
ValueError: if the ids are not of ... | [
"def",
"decode",
"(",
"self",
",",
"ids",
",",
"strip_extraneous",
"=",
"False",
")",
":",
"del",
"strip_extraneous",
"return",
"\" \"",
".",
"join",
"(",
"[",
"str",
"(",
"i",
")",
"for",
"i",
"in",
"ids",
"]",
")"
] | Transform sequence of float values into string (float values).
Args:
ids: array of floats to be converted.
strip_extraneous: unused
Returns:
String having space separated float values.
Raises:
ValueError: if the ids are not of the appropriate size. | [
"Transform",
"sequence",
"of",
"float",
"values",
"into",
"string",
"(",
"float",
"values",
")",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L1050-L1064 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | _pack_images | def _pack_images(images, rows, cols):
"""Helper utility to make a tiled field of images from numpy arrays.
Args:
images: Image tensor in shape [N, W, H, C].
rows: Number of images per row in tiled image.
cols: Number of images per column in tiled image.
Returns:
A tiled image of shape [W * rows,... | python | def _pack_images(images, rows, cols):
"""Helper utility to make a tiled field of images from numpy arrays.
Args:
images: Image tensor in shape [N, W, H, C].
rows: Number of images per row in tiled image.
cols: Number of images per column in tiled image.
Returns:
A tiled image of shape [W * rows,... | [
"def",
"_pack_images",
"(",
"images",
",",
"rows",
",",
"cols",
")",
":",
"shape",
"=",
"onp",
".",
"shape",
"(",
"images",
")",
"width",
",",
"height",
",",
"depth",
"=",
"shape",
"[",
"-",
"3",
":",
"]",
"images",
"=",
"onp",
".",
"reshape",
"(... | Helper utility to make a tiled field of images from numpy arrays.
Args:
images: Image tensor in shape [N, W, H, C].
rows: Number of images per row in tiled image.
cols: Number of images per column in tiled image.
Returns:
A tiled image of shape [W * rows, H * cols, C].
Truncates incomplete row... | [
"Helper",
"utility",
"to",
"make",
"a",
"tiled",
"field",
"of",
"images",
"from",
"numpy",
"arrays",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L49-L71 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | markdownify_operative_config_str | def markdownify_operative_config_str(string):
"""Convert an operative config string to markdown format."""
# TODO(b/37527917): Total hack below. Implement more principled formatting.
def process(line):
"""Convert a single line to markdown format."""
if not line.startswith('#'):
return ' ' + line... | python | def markdownify_operative_config_str(string):
"""Convert an operative config string to markdown format."""
# TODO(b/37527917): Total hack below. Implement more principled formatting.
def process(line):
"""Convert a single line to markdown format."""
if not line.startswith('#'):
return ' ' + line... | [
"def",
"markdownify_operative_config_str",
"(",
"string",
")",
":",
"# TODO(b/37527917): Total hack below. Implement more principled formatting.",
"def",
"process",
"(",
"line",
")",
":",
"\"\"\"Convert a single line to markdown format.\"\"\"",
"if",
"not",
"line",
".",
"startswi... | Convert an operative config string to markdown format. | [
"Convert",
"an",
"operative",
"config",
"string",
"to",
"markdown",
"format",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L326-L350 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.close | def close(self):
"""Close SummaryWriter. Final!"""
if not self._closed:
self._event_writer.close()
self._closed = True
del self._event_writer | python | def close(self):
"""Close SummaryWriter. Final!"""
if not self._closed:
self._event_writer.close()
self._closed = True
del self._event_writer | [
"def",
"close",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"_closed",
":",
"self",
".",
"_event_writer",
".",
"close",
"(",
")",
"self",
".",
"_closed",
"=",
"True",
"del",
"self",
".",
"_event_writer"
] | Close SummaryWriter. Final! | [
"Close",
"SummaryWriter",
".",
"Final!"
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L98-L103 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.scalar | def scalar(self, tag, value, step=None):
"""Saves scalar value.
Args:
tag: str: label for this data
value: int/float: number to log
step: int: training step
"""
value = float(onp.array(value))
if step is None:
step = self._step
else:
self._step = step
summary =... | python | def scalar(self, tag, value, step=None):
"""Saves scalar value.
Args:
tag: str: label for this data
value: int/float: number to log
step: int: training step
"""
value = float(onp.array(value))
if step is None:
step = self._step
else:
self._step = step
summary =... | [
"def",
"scalar",
"(",
"self",
",",
"tag",
",",
"value",
",",
"step",
"=",
"None",
")",
":",
"value",
"=",
"float",
"(",
"onp",
".",
"array",
"(",
"value",
")",
")",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
".",
"_step",
"else",
":"... | Saves scalar value.
Args:
tag: str: label for this data
value: int/float: number to log
step: int: training step | [
"Saves",
"scalar",
"value",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L111-L125 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.image | def image(self, tag, image, step=None):
"""Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3].
Args:
tag: str: label for this data
image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/
step: int: training step
"""
image = onp.array(image)
if s... | python | def image(self, tag, image, step=None):
"""Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3].
Args:
tag: str: label for this data
image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/
step: int: training step
"""
image = onp.array(image)
if s... | [
"def",
"image",
"(",
"self",
",",
"tag",
",",
"image",
",",
"step",
"=",
"None",
")",
":",
"image",
"=",
"onp",
".",
"array",
"(",
"image",
")",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
".",
"_step",
"else",
":",
"self",
".",
"_ste... | Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3].
Args:
tag: str: label for this data
image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/
step: int: training step | [
"Saves",
"RGB",
"image",
"summary",
"from",
"onp",
".",
"ndarray",
"[",
"H",
"W",
"]",
"[",
"H",
"W",
"1",
"]",
"or",
"[",
"H",
"W",
"3",
"]",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L127-L152 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.images | def images(self, tag, images, step=None, rows=None, cols=None):
"""Saves (rows, cols) tiled images from onp.ndarray.
If either rows or cols aren't given, they are determined automatically
from the size of the image batch, if neither are given a long column
of images is produced. This truncates the imag... | python | def images(self, tag, images, step=None, rows=None, cols=None):
"""Saves (rows, cols) tiled images from onp.ndarray.
If either rows or cols aren't given, they are determined automatically
from the size of the image batch, if neither are given a long column
of images is produced. This truncates the imag... | [
"def",
"images",
"(",
"self",
",",
"tag",
",",
"images",
",",
"step",
"=",
"None",
",",
"rows",
"=",
"None",
",",
"cols",
"=",
"None",
")",
":",
"images",
"=",
"onp",
".",
"array",
"(",
"images",
")",
"if",
"step",
"is",
"None",
":",
"step",
"=... | Saves (rows, cols) tiled images from onp.ndarray.
If either rows or cols aren't given, they are determined automatically
from the size of the image batch, if neither are given a long column
of images is produced. This truncates the image batch rather than padding
if it doesn't fill the final row.
... | [
"Saves",
"(",
"rows",
"cols",
")",
"tiled",
"images",
"from",
"onp",
".",
"ndarray",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L154-L183 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.plot | def plot(self, tag, mpl_plt, step=None, close_plot=True):
"""Saves matplotlib plot output to summary image.
Args:
tag: str: label for this data
mpl_plt: matplotlib stateful pyplot object with prepared plotting state
step: int: training step
close_plot: bool: automatically closes plot
... | python | def plot(self, tag, mpl_plt, step=None, close_plot=True):
"""Saves matplotlib plot output to summary image.
Args:
tag: str: label for this data
mpl_plt: matplotlib stateful pyplot object with prepared plotting state
step: int: training step
close_plot: bool: automatically closes plot
... | [
"def",
"plot",
"(",
"self",
",",
"tag",
",",
"mpl_plt",
",",
"step",
"=",
"None",
",",
"close_plot",
"=",
"True",
")",
":",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
".",
"_step",
"else",
":",
"self",
".",
"_step",
"=",
"step",
"fig",... | Saves matplotlib plot output to summary image.
Args:
tag: str: label for this data
mpl_plt: matplotlib stateful pyplot object with prepared plotting state
step: int: training step
close_plot: bool: automatically closes plot | [
"Saves",
"matplotlib",
"plot",
"output",
"to",
"summary",
"image",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L185-L210 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.audio | def audio(self, tag, audiodata, step=None, sample_rate=44100):
"""Saves audio.
NB: single channel only right now.
Args:
tag: str: label for this data
audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave
step: int: training step
sample_rate: sample rate of passed ... | python | def audio(self, tag, audiodata, step=None, sample_rate=44100):
"""Saves audio.
NB: single channel only right now.
Args:
tag: str: label for this data
audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave
step: int: training step
sample_rate: sample rate of passed ... | [
"def",
"audio",
"(",
"self",
",",
"tag",
",",
"audiodata",
",",
"step",
"=",
"None",
",",
"sample_rate",
"=",
"44100",
")",
":",
"audiodata",
"=",
"onp",
".",
"array",
"(",
"audiodata",
")",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
"."... | Saves audio.
NB: single channel only right now.
Args:
tag: str: label for this data
audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave
step: int: training step
sample_rate: sample rate of passed in audio buffer | [
"Saves",
"audio",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L212-L249 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.histogram | def histogram(self, tag, values, bins, step=None):
"""Saves histogram of values.
Args:
tag: str: label for this data
values: ndarray: will be flattened by this routine
bins: number of bins in histogram, or array of bins for onp.histogram
step: int: training step
"""
if step is N... | python | def histogram(self, tag, values, bins, step=None):
"""Saves histogram of values.
Args:
tag: str: label for this data
values: ndarray: will be flattened by this routine
bins: number of bins in histogram, or array of bins for onp.histogram
step: int: training step
"""
if step is N... | [
"def",
"histogram",
"(",
"self",
",",
"tag",
",",
"values",
",",
"bins",
",",
"step",
"=",
"None",
")",
":",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
".",
"_step",
"else",
":",
"self",
".",
"_step",
"=",
"step",
"values",
"=",
"onp",... | Saves histogram of values.
Args:
tag: str: label for this data
values: ndarray: will be flattened by this routine
bins: number of bins in histogram, or array of bins for onp.histogram
step: int: training step | [
"Saves",
"histogram",
"of",
"values",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L251-L287 | train |
tensorflow/tensor2tensor | tensor2tensor/trax/jaxboard.py | SummaryWriter.text | def text(self, tag, textdata, step=None):
"""Saves a text summary.
Args:
tag: str: label for this data
textdata: string, or 1D/2D list/numpy array of strings
step: int: training step
Note: markdown formatting is rendered by tensorboard.
"""
if step is None:
step = self._step... | python | def text(self, tag, textdata, step=None):
"""Saves a text summary.
Args:
tag: str: label for this data
textdata: string, or 1D/2D list/numpy array of strings
step: int: training step
Note: markdown formatting is rendered by tensorboard.
"""
if step is None:
step = self._step... | [
"def",
"text",
"(",
"self",
",",
"tag",
",",
"textdata",
",",
"step",
"=",
"None",
")",
":",
"if",
"step",
"is",
"None",
":",
"step",
"=",
"self",
".",
"_step",
"else",
":",
"self",
".",
"_step",
"=",
"step",
"smd",
"=",
"SummaryMetadata",
"(",
"... | Saves a text summary.
Args:
tag: str: label for this data
textdata: string, or 1D/2D list/numpy array of strings
step: int: training step
Note: markdown formatting is rendered by tensorboard. | [
"Saves",
"a",
"text",
"summary",
"."
] | 272500b6efe353aeb638d2745ed56e519462ca31 | https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L289-L322 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.