repo
stringlengths
7
55
path
stringlengths
4
223
func_name
stringlengths
1
134
original_string
stringlengths
75
104k
language
stringclasses
1 value
code
stringlengths
75
104k
code_tokens
listlengths
19
28.4k
docstring
stringlengths
1
46.9k
docstring_tokens
listlengths
1
1.97k
sha
stringlengths
40
40
url
stringlengths
87
315
partition
stringclasses
1 value
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ResidualFeedForward
def ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode): """Residual feed-forward layer with normalization at start.""" return layers.Residual( layers.LayerNorm(), layers.Dense(feedforward_depth), layers.Relu(...
python
def ResidualFeedForward(feature_depth, feedforward_depth, dropout, mode): """Residual feed-forward layer with normalization at start.""" return layers.Residual( layers.LayerNorm(), layers.Dense(feedforward_depth), layers.Relu(...
[ "def", "ResidualFeedForward", "(", "feature_depth", ",", "feedforward_depth", ",", "dropout", ",", "mode", ")", ":", "return", "layers", ".", "Residual", "(", "layers", ".", "LayerNorm", "(", ")", ",", "layers", ".", "Dense", "(", "feedforward_depth", ")", "...
Residual feed-forward layer with normalization at start.
[ "Residual", "feed", "-", "forward", "layer", "with", "normalization", "at", "start", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L24-L36
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
EncoderLayer
def EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to t...
python
def EncoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to t...
[ "def", "EncoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "mode", ")", ":", "# The encoder block expects (activation, mask) as input and returns", "# the new activations only, we add the mask back to output next.", "encoder_block", ...
Transformer encoder layer. The input to the encoder is a pair (embedded source, mask) where the mask is created from the original source to prevent attending to the padding part of the input. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_head...
[ "Transformer", "encoder", "layer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L39-L76
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
TransformerEncoder
def TransformerEncoder(vocab_size, num_classes=10, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, ...
python
def TransformerEncoder(vocab_size, num_classes=10, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, ...
[ "def", "TransformerEncoder", "(", "vocab_size", ",", "num_classes", "=", "10", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_layers", "=", "6", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1", ",", "max_len", "="...
Transformer encoder. Args: vocab_size: int: vocab size num_classes: how many classes on output feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dropout: f...
[ "Transformer", "encoder", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L79-L120
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
DecoderLayer
def DecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of att...
python
def DecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, mode): """Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of att...
[ "def", "DecoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "mode", ")", ":", "return", "layers", ".", "Serial", "(", "layers", ".", "Residual", "(", "# Self-attention block.", "layers", ".", "LayerNorm", "(", ...
Transformer decoder layer. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) mode: str: 'train' or 'eval' Returns: the layer.
[ "Transformer", "decoder", "layer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L123-L151
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
TransformerLM
def TransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, mode='train'): """Transformer language model (only uses the decod...
python
def TransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, max_len=2048, mode='train'): """Transformer language model (only uses the decod...
[ "def", "TransformerLM", "(", "vocab_size", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_layers", "=", "6", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1", ",", "max_len", "=", "2048", ",", "mode", "=", "'trai...
Transformer language model (only uses the decoder part of Transformer). Args: vocab_size: int: vocab size feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_layers: int: number of encoder/decoder layers num_heads: int: number of attention heads dro...
[ "Transformer", "language", "model", "(", "only", "uses", "the", "decoder", "part", "of", "Transformer", ")", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L154-L188
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ChunkedDecoderLayer
def ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode): """Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embe...
python
def ChunkedDecoderLayer(feature_depth, feedforward_depth, num_heads, dropout, chunk_selector, mode): """Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embe...
[ "def", "ChunkedDecoderLayer", "(", "feature_depth", ",", "feedforward_depth", ",", "num_heads", ",", "dropout", ",", "chunk_selector", ",", "mode", ")", ":", "return", "layers", ".", "Serial", "(", "layers", ".", "Residual", "(", "# Self-attention block.", "layers...
Transformer decoder layer operating on chunks. Args: feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads: int: number of attention heads dropout: float: dropout rate (how much to drop out) chunk_selector: a function from chunk number to list of ch...
[ "Transformer", "decoder", "layer", "operating", "on", "chunks", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L191-L220
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
ChunkedTransformerLM
def ChunkedTransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, chunk_selector=None, max_...
python
def ChunkedTransformerLM(vocab_size, feature_depth=512, feedforward_depth=2048, num_layers=6, num_heads=8, dropout=0.1, chunk_selector=None, max_...
[ "def", "ChunkedTransformerLM", "(", "vocab_size", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_layers", "=", "6", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1", ",", "chunk_selector", "=", "None", ",", "max_len"...
Transformer language model operating on chunks. The input to this model is a sequence presented as a list or tuple of chunks: (chunk1, chunk2, chunks3, ..., chunkN). Each chunk should have the same shape (batch, chunk-length) and together they represent a long sequence that's a concatenation chunk1,chunk2,....
[ "Transformer", "language", "model", "operating", "on", "chunks", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L223-L273
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/transformer.py
Transformer
def Transformer(source_vocab_size, target_vocab_size, mode='train', num_layers=6, feature_depth=512, feedforward_depth=2048, num_heads=8, dropout=0.1, shared_embedding=True, ma...
python
def Transformer(source_vocab_size, target_vocab_size, mode='train', num_layers=6, feature_depth=512, feedforward_depth=2048, num_heads=8, dropout=0.1, shared_embedding=True, ma...
[ "def", "Transformer", "(", "source_vocab_size", ",", "target_vocab_size", ",", "mode", "=", "'train'", ",", "num_layers", "=", "6", ",", "feature_depth", "=", "512", ",", "feedforward_depth", "=", "2048", ",", "num_heads", "=", "8", ",", "dropout", "=", "0.1...
Transformer model. Args: source_vocab_size: int: source vocab size target_vocab_size: int: target vocab size mode: str: 'train' or 'eval' num_layers: int: number of encoder/decoder layers feature_depth: int: depth of embedding feedforward_depth: int: depth of feed-forward layer num_heads...
[ "Transformer", "model", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/transformer.py#L279-L431
train
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_base
def mtf_transformer_base(): """Set of hyperparameters.""" hparams = common_hparams.basic_params1() hparams.no_data_parallelism = True hparams.use_fixed_batch_size = True hparams.add_hparam("mtf_mode", True) hparams.batch_size = 64 hparams.max_length = 256 hparams.add_hparam("d_model", 512) hparams.add...
python
def mtf_transformer_base(): """Set of hyperparameters.""" hparams = common_hparams.basic_params1() hparams.no_data_parallelism = True hparams.use_fixed_batch_size = True hparams.add_hparam("mtf_mode", True) hparams.batch_size = 64 hparams.max_length = 256 hparams.add_hparam("d_model", 512) hparams.add...
[ "def", "mtf_transformer_base", "(", ")", ":", "hparams", "=", "common_hparams", ".", "basic_params1", "(", ")", "hparams", ".", "no_data_parallelism", "=", "True", "hparams", ".", "use_fixed_batch_size", "=", "True", "hparams", ".", "add_hparam", "(", "\"mtf_mode\...
Set of hyperparameters.
[ "Set", "of", "hyperparameters", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L791-L883
train
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_tiny
def mtf_transformer_tiny(): """Catch bugs locally...""" hparams = mtf_transformer_base() hparams.d_model = 128 hparams.d_ff = 512 hparams.batch_size = 8 hparams.encoder_layers = ["att", "drd"] * 2 hparams.decoder_layers = ["att", "enc_att", "drd"] * 2 hparams.num_heads = 8 # data parallelism and model...
python
def mtf_transformer_tiny(): """Catch bugs locally...""" hparams = mtf_transformer_base() hparams.d_model = 128 hparams.d_ff = 512 hparams.batch_size = 8 hparams.encoder_layers = ["att", "drd"] * 2 hparams.decoder_layers = ["att", "enc_att", "drd"] * 2 hparams.num_heads = 8 # data parallelism and model...
[ "def", "mtf_transformer_tiny", "(", ")", ":", "hparams", "=", "mtf_transformer_base", "(", ")", "hparams", ".", "d_model", "=", "128", "hparams", ".", "d_ff", "=", "512", "hparams", ".", "batch_size", "=", "8", "hparams", ".", "encoder_layers", "=", "[", "...
Catch bugs locally...
[ "Catch", "bugs", "locally", "..." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L897-L909
train
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_paper_lm
def mtf_transformer_paper_lm(size): """Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles e...
python
def mtf_transformer_paper_lm(size): """Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles e...
[ "def", "mtf_transformer_paper_lm", "(", "size", ")", ":", "n", "=", "2", "**", "size", "hparams", "=", "mtf_transformer_base_lm", "(", ")", "hparams", ".", "batch_size", "=", "256", "hparams", ".", "d_model", "=", "1024", "hparams", ".", "d_ff", "=", "int"...
Config for language-model experiments. Train these on languagemodel_lm1b32k_packed for 136000 steps (10 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Results: size params...
[ "Config", "for", "language", "-", "model", "experiments", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L953-L989
train
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_paper_tr
def mtf_transformer_paper_tr(size): """Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each ...
python
def mtf_transformer_paper_tr(size): """Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each ...
[ "def", "mtf_transformer_paper_tr", "(", "size", ")", ":", "n", "=", "2", "**", "size", "hparams", "=", "mtf_transformer_base", "(", ")", "hparams", ".", "label_smoothing", "=", "0.1", "hparams", ".", "batch_size", "=", "128", "hparams", ".", "d_model", "=", ...
Config for translation experiments. Train these on translate_enfr_wmt32k_packed for 154000 steps (3 epochs) The size parameter is an integer that controls the number of heads and the size of the size of the feedforward hidden layers. Increasing size by 1 doubles each of these. Args: size: an integer ...
[ "Config", "for", "translation", "experiments", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1041-L1065
train
tensorflow/tensor2tensor
tensor2tensor/models/mtf_transformer.py
mtf_transformer_lm_baseline
def mtf_transformer_lm_baseline(): """Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams """ hparams = mtf_transformer_paper_lm(-1) hparams...
python
def mtf_transformer_lm_baseline(): """Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams """ hparams = mtf_transformer_paper_lm(-1) hparams...
[ "def", "mtf_transformer_lm_baseline", "(", ")", ":", "hparams", "=", "mtf_transformer_paper_lm", "(", "-", "1", ")", "hparams", ".", "batch_size", "=", "128", "hparams", ".", "learning_rate_decay_steps", "=", "27200", "# one epoch on lm1b", "hparams", ".", "mesh_sha...
Small language model to run on 1 TPU. Run this on 2x2 on languagemodel_lm1b32k_packed for 272000 steps (10 epochs) Results: params/10^9 log-ppl(per-token) 0.14 3.202 Returns: a hparams
[ "Small", "language", "model", "to", "run", "on", "1", "TPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/mtf_transformer.py#L1171-L1186
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
multihead_graph_attention
def multihead_graph_attention(query_antecedent, memory_antecedent, bias, total_key_depth, total_value_depth, output_depth, num_heads, ...
python
def multihead_graph_attention(query_antecedent, memory_antecedent, bias, total_key_depth, total_value_depth, output_depth, num_heads, ...
[ "def", "multihead_graph_attention", "(", "query_antecedent", ",", "memory_antecedent", ",", "bias", ",", "total_key_depth", ",", "total_value_depth", ",", "output_depth", ",", "num_heads", ",", "dropout_rate", ",", "image_shapes", "=", "None", ",", "attention_type", "...
Multihead scaled-dot-product attention with input/output transformations. Args: query_antecedent: a Tensor with shape [batch, length_q, channels] memory_antecedent: a Tensor with shape [batch, length_m, channels] or None bias: bias Tensor (see attention_bias()) total_key_depth: an integer total_v...
[ "Multihead", "scaled", "-", "dot", "-", "product", "attention", "with", "input", "/", "output", "transformations", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L28-L146
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
graph_attention
def graph_attention(q, k, v, bias, dropout_rate=0.0, image_shapes=None, name=None, make_image_summary=True, save_weights_to=None, dropout_br...
python
def graph_attention(q, k, v, bias, dropout_rate=0.0, image_shapes=None, name=None, make_image_summary=True, save_weights_to=None, dropout_br...
[ "def", "graph_attention", "(", "q", ",", "k", ",", "v", ",", "bias", ",", "dropout_rate", "=", "0.0", ",", "image_shapes", "=", "None", ",", "name", "=", "None", ",", "make_image_summary", "=", "True", ",", "save_weights_to", "=", "None", ",", "dropout_b...
graph attention. Args: q: a Tensor with shape [batch, heads, length_q, depth_k] k: a Tensor with shape [batch, heads, length_kv, depth_k] v: a Tensor with shape [batch, heads, length_kv, depth_v] bias: bias Tensor (see attention_bias()) dropout_rate: a floating point number image_shapes: opti...
[ "graph", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L185-L248
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
_compute_edge_transforms
def _compute_edge_transforms(node_states, depth, num_transforms, name="transform"): """Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph...
python
def _compute_edge_transforms(node_states, depth, num_transforms, name="transform"): """Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph...
[ "def", "_compute_edge_transforms", "(", "node_states", ",", "depth", ",", "num_transforms", ",", "name", "=", "\"transform\"", ")", ":", "node_shapes", "=", "common_layers", ".", "shape_list", "(", "node_states", ")", "x", "=", "common_layers", ".", "dense", "("...
Helper function that computes transformation for keys and values. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_d...
[ "Helper", "function", "that", "computes", "transformation", "for", "keys", "and", "values", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L251-L301
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
compute_mpnn_qkv
def compute_mpnn_qkv(node_states, total_key_depth, total_value_depth, num_transforms): """Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidd...
python
def compute_mpnn_qkv(node_states, total_key_depth, total_value_depth, num_transforms): """Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidd...
[ "def", "compute_mpnn_qkv", "(", "node_states", ",", "total_key_depth", ",", "total_value_depth", ",", "num_transforms", ")", ":", "# node_states is initially a tensor with shape [B, N, D]. The call to dense", "# creates a D x K kernel that serves as a fully-connected layer.", "#", "# F...
Computes query, key and value for edge matrices. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total_value_depth). Let T be...
[ "Computes", "query", "key", "and", "value", "for", "edge", "matrices", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L304-L364
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
sparse_message_pass_batched
def sparse_message_pass_batched(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, ...
python
def sparse_message_pass_batched(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, ...
[ "def", "sparse_message_pass_batched", "(", "node_states", ",", "adjacency_matrices", ",", "num_edge_types", ",", "hidden_size", ",", "use_bias", "=", "True", ",", "average_aggregation", "=", "False", ",", "name", "=", "\"sparse_ggnn_batched\"", ")", ":", "b", ",", ...
Identical to sparse_ggnn except that each input has a batch dimension. B = The batch size. N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial states of each node in the graph. Shape: [B, N, H] adjacency_matrices: Adjacen...
[ "Identical", "to", "sparse_ggnn", "except", "that", "each", "input", "has", "a", "batch", "dimension", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L367-L428
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
sparse_message_pass
def sparse_message_pass(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn"): """One message-passing st...
python
def sparse_message_pass(node_states, adjacency_matrices, num_edge_types, hidden_size, use_bias=True, average_aggregation=False, name="sparse_ggnn"): """One message-passing st...
[ "def", "sparse_message_pass", "(", "node_states", ",", "adjacency_matrices", ",", "num_edge_types", ",", "hidden_size", ",", "use_bias", "=", "True", ",", "average_aggregation", "=", "False", ",", "name", "=", "\"sparse_ggnn\"", ")", ":", "n", "=", "tf", ".", ...
One message-passing step for a GNN with a sparse adjacency matrix. Implements equation 2 (the message passing step) in [Li et al. 2015](https://arxiv.org/abs/1511.05493). N = The number of nodes in each batch. H = The size of the hidden states. T = The number of edge types. Args: node_states: Initial...
[ "One", "message", "-", "passing", "step", "for", "a", "GNN", "with", "a", "sparse", "adjacency", "matrix", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L431-L511
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
multihead_mpnn_attention
def multihead_mpnn_attention(node_states, total_key_depth, total_value_depth, output_depth, num_heads, adjacency_matrix=None, num_edge_types=5, ...
python
def multihead_mpnn_attention(node_states, total_key_depth, total_value_depth, output_depth, num_heads, adjacency_matrix=None, num_edge_types=5, ...
[ "def", "multihead_mpnn_attention", "(", "node_states", ",", "total_key_depth", ",", "total_value_depth", ",", "output_depth", ",", "num_heads", ",", "adjacency_matrix", "=", "None", ",", "num_edge_types", "=", "5", ",", "num_transforms", "=", "None", ",", "use_weigh...
Multihead scaled-dot-product attention with input/output transformations. Let B be the number of batches. Let N be the number of nodes in the graph. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries (total_key_depth). Let V be the size of the attention values (total...
[ "Multihead", "scaled", "-", "dot", "-", "product", "attention", "with", "input", "/", "output", "transformations", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L514-L649
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
dot_product_mpnn_attention
def dot_product_mpnn_attention(q, k, v, adjacency_matrix, num_edge_types, num_transforms=None, use_weighted_sum=False, ...
python
def dot_product_mpnn_attention(q, k, v, adjacency_matrix, num_edge_types, num_transforms=None, use_weighted_sum=False, ...
[ "def", "dot_product_mpnn_attention", "(", "q", ",", "k", ",", "v", ",", "adjacency_matrix", ",", "num_edge_types", ",", "num_transforms", "=", "None", ",", "use_weighted_sum", "=", "False", ",", "name", "=", "None", ")", ":", "with", "tf", ".", "variable_sco...
Dot product attention with edge vectors. Let B be the number of batches. Let N be the number of nodes in the graph. Let K be the size of the attention keys/queries. Let V be the size of the attention values. Let T be the total number of transforms (num_transforms). Args: q: The query Tensor of shape [...
[ "Dot", "product", "attention", "with", "edge", "vectors", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L652-L783
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
ggnn_fast_dense
def ggnn_fast_dense(node_states, adjacency_matrix, num_edge_types, total_value_depth, name=None): """ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be ...
python
def ggnn_fast_dense(node_states, adjacency_matrix, num_edge_types, total_value_depth, name=None): """ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be ...
[ "def", "ggnn_fast_dense", "(", "node_states", ",", "adjacency_matrix", ",", "num_edge_types", ",", "total_value_depth", ",", "name", "=", "None", ")", ":", "# between the same nodes (with only one edge of each type. adjacency_matrix", "# will need to be converted to shape [B, T, N,...
ggnn version of the MPNN from Gilmer et al. Let B be the number of batches. Let D be the size of the node hidden states. Let K be the size of the attention keys/queries. Let V be the size of the output of the ggnn. Let T be the number of transforms / edge types. Args: node_states: The value Tensor of ...
[ "ggnn", "version", "of", "the", "MPNN", "from", "Gilmer", "et", "al", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L786-L837
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
compute_values
def compute_values(edge_compatibility, v): """Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, dept...
python
def compute_values(edge_compatibility, v): """Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, dept...
[ "def", "compute_values", "(", "edge_compatibility", ",", "v", ")", ":", "# Computes the incoming value vectors for each node by weighting them", "# according to the attention weights. These values are still segregated by", "# edge type.", "# Shape = [B, T, N, V].", "all_edge_values", "=", ...
Compute values. If edge compatibilities is just adjacency, we get ggnn. Args: edge_compatibility: A tensor of shape [batch, num_transforms, length, depth] v: A tensor of shape [batch, num_transforms, length, depth] Returns: output: A [batch, length, depth] tensor
[ "Compute", "values", ".", "If", "edge", "compatibilities", "is", "just", "adjacency", "we", "get", "ggnn", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L840-L860
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
precompute_edge_matrices
def precompute_edge_matrices(adjacency, hparams): """Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] ...
python
def precompute_edge_matrices(adjacency, hparams): """Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] ...
[ "def", "precompute_edge_matrices", "(", "adjacency", ",", "hparams", ")", ":", "batch_size", ",", "num_nodes", ",", "_", ",", "edge_dim", "=", "common_layers", ".", "shape_list", "(", "adjacency", ")", "# build the edge_network for incoming edges", "with", "tf", "."...
Precompute the a_in and a_out tensors. (we don't want to add to the graph everytime _fprop is called) Args: adjacency: placeholder of real valued vectors of shape [B, L, L, E] hparams: HParams object Returns: edge_matrices: [batch, L * D, L * D] the dense matrix for message passing viewed as a bl...
[ "Precompute", "the", "a_in", "and", "a_out", "tensors", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L863-L907
train
tensorflow/tensor2tensor
tensor2tensor/layers/message_passing_attention.py
dense_message_pass
def dense_message_pass(node_states, edge_matrices): """Computes a_t from h_{t-1}, see bottom of page 3 in the paper. Args: node_states: [B, L, D] tensor (h_{t-1}) edge_matrices (tf.float32): [B, L*D, L*D] Returns: messages (tf.float32): [B, L, D] For each pair of nodes in the graph a message i...
python
def dense_message_pass(node_states, edge_matrices): """Computes a_t from h_{t-1}, see bottom of page 3 in the paper. Args: node_states: [B, L, D] tensor (h_{t-1}) edge_matrices (tf.float32): [B, L*D, L*D] Returns: messages (tf.float32): [B, L, D] For each pair of nodes in the graph a message i...
[ "def", "dense_message_pass", "(", "node_states", ",", "edge_matrices", ")", ":", "batch_size", ",", "num_nodes", ",", "node_dim", "=", "common_layers", ".", "shape_list", "(", "node_states", ")", "# Stack the nodes as a big column vector.", "h_flat", "=", "tf", ".", ...
Computes a_t from h_{t-1}, see bottom of page 3 in the paper. Args: node_states: [B, L, D] tensor (h_{t-1}) edge_matrices (tf.float32): [B, L*D, L*D] Returns: messages (tf.float32): [B, L, D] For each pair of nodes in the graph a message is sent along both the incoming and outgoing edge.
[ "Computes", "a_t", "from", "h_", "{", "t", "-", "1", "}", "see", "bottom", "of", "page", "3", "in", "the", "paper", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/layers/message_passing_attention.py#L910-L935
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
to_example
def to_example(dictionary): """Helper: build tf.Example from (string -> int/float/str list) dictionary.""" features = {} for (k, v) in six.iteritems(dictionary): if not v: raise ValueError("Empty generated field: %s" % str((k, v))) if isinstance(v[0], six.integer_types): features[k] = tf.train...
python
def to_example(dictionary): """Helper: build tf.Example from (string -> int/float/str list) dictionary.""" features = {} for (k, v) in six.iteritems(dictionary): if not v: raise ValueError("Empty generated field: %s" % str((k, v))) if isinstance(v[0], six.integer_types): features[k] = tf.train...
[ "def", "to_example", "(", "dictionary", ")", ":", "features", "=", "{", "}", "for", "(", "k", ",", "v", ")", "in", "six", ".", "iteritems", "(", "dictionary", ")", ":", "if", "not", "v", ":", "raise", "ValueError", "(", "\"Empty generated field: %s\"", ...
Helper: build tf.Example from (string -> int/float/str list) dictionary.
[ "Helper", ":", "build", "tf", ".", "Example", "from", "(", "string", "-", ">", "int", "/", "float", "/", "str", "list", ")", "dictionary", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L43-L62
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_files_distributed
def generate_files_distributed(generator, output_name, output_dir, num_shards=1, max_cases=None, task_id=0): """generate_files but with a single writer writing to ...
python
def generate_files_distributed(generator, output_name, output_dir, num_shards=1, max_cases=None, task_id=0): """generate_files but with a single writer writing to ...
[ "def", "generate_files_distributed", "(", "generator", ",", "output_name", ",", "output_dir", ",", "num_shards", "=", "1", ",", "max_cases", "=", "None", ",", "task_id", "=", "0", ")", ":", "assert", "task_id", "<", "num_shards", "output_filename", "=", "shard...
generate_files but with a single writer writing to shard task_id.
[ "generate_files", "but", "with", "a", "single", "writer", "writing", "to", "shard", "task_id", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L65-L89
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_files
def generate_files(generator, output_filenames, max_cases=None, cycle_every_n=1): """Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. ...
python
def generate_files(generator, output_filenames, max_cases=None, cycle_every_n=1): """Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. ...
[ "def", "generate_files", "(", "generator", ",", "output_filenames", ",", "max_cases", "=", "None", ",", "cycle_every_n", "=", "1", ")", ":", "if", "outputs_exist", "(", "output_filenames", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Skipping generator...
Generate cases from a generator and save as TFRecord files. Generated cases are transformed to tf.Example protos and saved as TFRecords in sharded files named output_dir/output_name-00..N-of-00..M=num_shards. Args: generator: a generator yielding (string -> int/float/str list) dictionaries. output_filen...
[ "Generate", "cases", "from", "a", "generator", "and", "save", "as", "TFRecord", "files", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L134-L193
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
download_report_hook
def download_report_hook(count, block_size, total_size): """Report hook for download progress. Args: count: current block number block_size: block size total_size: total size """ percent = int(count * block_size * 100 / total_size) print("\r%d%%" % percent + " completed", end="\r")
python
def download_report_hook(count, block_size, total_size): """Report hook for download progress. Args: count: current block number block_size: block size total_size: total size """ percent = int(count * block_size * 100 / total_size) print("\r%d%%" % percent + " completed", end="\r")
[ "def", "download_report_hook", "(", "count", ",", "block_size", ",", "total_size", ")", ":", "percent", "=", "int", "(", "count", "*", "block_size", "*", "100", "/", "total_size", ")", "print", "(", "\"\\r%d%%\"", "%", "percent", "+", "\" completed\"", ",", ...
Report hook for download progress. Args: count: current block number block_size: block size total_size: total size
[ "Report", "hook", "for", "download", "progress", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L196-L205
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
maybe_download
def maybe_download(directory, filename, uri): """Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filen...
python
def maybe_download(directory, filename, uri): """Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filen...
[ "def", "maybe_download", "(", "directory", ",", "filename", ",", "uri", ")", ":", "tf", ".", "gfile", ".", "MakeDirs", "(", "directory", ")", "filepath", "=", "os", ".", "path", ".", "join", "(", "directory", ",", "filename", ")", "if", "tf", ".", "g...
Download filename from uri unless it's already in directory. Copies a remote file to local if that local file does not already exist. If the local file pre-exists this function call, it does not check that the local file is a copy of the remote. Remote filenames can be filepaths, any URI readable by tensorfl...
[ "Download", "filename", "from", "uri", "unless", "it", "s", "already", "in", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L208-L248
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
maybe_download_from_drive
def maybe_download_from_drive(directory, filename, url): """Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns:...
python
def maybe_download_from_drive(directory, filename, url): """Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns:...
[ "def", "maybe_download_from_drive", "(", "directory", ",", "filename", ",", "url", ")", ":", "if", "not", "tf", ".", "gfile", ".", "Exists", "(", "directory", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Creating directory %s\"", "%", "directory", ...
Download filename from Google drive unless it's already in directory. Args: directory: path to the directory that will be used. filename: name of the file to download to (do nothing if it already exists). url: URL to download from. Returns: The path to the downloaded file.
[ "Download", "filename", "from", "Google", "drive", "unless", "it", "s", "already", "in", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L251-L298
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
gunzip_file
def gunzip_file(gz_path, new_path): """Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped. """ if tf.gfile.Exists(new_path): tf.logging.info("File %s already exists, skipping unpacking" % new_path) return tf.logging.inf...
python
def gunzip_file(gz_path, new_path): """Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped. """ if tf.gfile.Exists(new_path): tf.logging.info("File %s already exists, skipping unpacking" % new_path) return tf.logging.inf...
[ "def", "gunzip_file", "(", "gz_path", ",", "new_path", ")", ":", "if", "tf", ".", "gfile", ".", "Exists", "(", "new_path", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"File %s already exists, skipping unpacking\"", "%", "new_path", ")", "return", "tf...
Unzips from gz_path into new_path. Args: gz_path: path to the zipped file. new_path: path to where the file will be unzipped.
[ "Unzips", "from", "gz_path", "into", "new_path", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L301-L318
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_vocab_inner
def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generator, max_subtoken_length=None, reserved_tokens=None): """Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are store...
python
def get_or_generate_vocab_inner(data_dir, vocab_filename, vocab_size, generator, max_subtoken_length=None, reserved_tokens=None): """Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are store...
[ "def", "get_or_generate_vocab_inner", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "generator", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ")", ":", "if", "data_dir", "and", "vocab_filename", ":", "vocab_filepath...
Inner implementation for vocab generators. Args: data_dir: The base directory where data and vocab files are stored. If None, then do not save the vocab even if it doesn't exist. vocab_filename: relative filename where vocab file is stored vocab_size: target size of the vocabulary constructed by Su...
[ "Inner", "implementation", "for", "vocab", "generators", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L321-L358
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_vocab
def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size, sources, file_byte_budget=1e6, max_subtoken_length=None): """Generate a vocabulary from the datasets in sources.""" vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_bud...
python
def get_or_generate_vocab(data_dir, tmp_dir, vocab_filename, vocab_size, sources, file_byte_budget=1e6, max_subtoken_length=None): """Generate a vocabulary from the datasets in sources.""" vocab_generator = generate_lines_for_vocab(tmp_dir, sources, file_byte_bud...
[ "def", "get_or_generate_vocab", "(", "data_dir", ",", "tmp_dir", ",", "vocab_filename", ",", "vocab_size", ",", "sources", ",", "file_byte_budget", "=", "1e6", ",", "max_subtoken_length", "=", "None", ")", ":", "vocab_generator", "=", "generate_lines_for_vocab", "("...
Generate a vocabulary from the datasets in sources.
[ "Generate", "a", "vocabulary", "from", "the", "datasets", "in", "sources", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L361-L368
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
generate_lines_for_vocab
def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6): """Generate lines for vocabulary generation.""" tf.logging.info("Generating vocab from: %s", str(sources)) for source in sources: url = source[0] filename = os.path.basename(url) compressed_file = maybe_download(tmp_dir, filename, u...
python
def generate_lines_for_vocab(tmp_dir, sources, file_byte_budget=1e6): """Generate lines for vocabulary generation.""" tf.logging.info("Generating vocab from: %s", str(sources)) for source in sources: url = source[0] filename = os.path.basename(url) compressed_file = maybe_download(tmp_dir, filename, u...
[ "def", "generate_lines_for_vocab", "(", "tmp_dir", ",", "sources", ",", "file_byte_budget", "=", "1e6", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Generating vocab from: %s\"", ",", "str", "(", "sources", ")", ")", "for", "source", "in", "sources", ...
Generate lines for vocabulary generation.
[ "Generate", "lines", "for", "vocabulary", "generation", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L371-L413
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_tabbed_vocab
def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename, index, vocab_filename, vocab_size): r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a...
python
def get_or_generate_tabbed_vocab(data_dir, tmp_dir, source_filename, index, vocab_filename, vocab_size): r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a...
[ "def", "get_or_generate_tabbed_vocab", "(", "data_dir", ",", "tmp_dir", ",", "source_filename", ",", "index", ",", "vocab_filename", ",", "vocab_size", ")", ":", "def", "generate", "(", ")", ":", "filepath", "=", "os", ".", "path", ".", "join", "(", "tmp_dir...
r"""Generate a vocabulary from a tabbed source file. The source is a file of source, target pairs, where each line contains a source string and a target string, separated by a tab ('\t') character. The index parameter specifies 0 for the source or 1 for the target. Args: data_dir: path to the data directo...
[ "r", "Generate", "a", "vocabulary", "from", "a", "tabbed", "source", "file", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L416-L447
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
get_or_generate_txt_vocab
def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size, filepatterns): """Generate a vocabulary from txt files with example-per-line.""" if isinstance(filepatterns, str): filepatterns = [filepatterns] def generate(): tf.logging.info("Generating vocab from %s", fi...
python
def get_or_generate_txt_vocab(data_dir, vocab_filename, vocab_size, filepatterns): """Generate a vocabulary from txt files with example-per-line.""" if isinstance(filepatterns, str): filepatterns = [filepatterns] def generate(): tf.logging.info("Generating vocab from %s", fi...
[ "def", "get_or_generate_txt_vocab", "(", "data_dir", ",", "vocab_filename", ",", "vocab_size", ",", "filepatterns", ")", ":", "if", "isinstance", "(", "filepatterns", ",", "str", ")", ":", "filepatterns", "=", "[", "filepatterns", "]", "def", "generate", "(", ...
Generate a vocabulary from txt files with example-per-line.
[ "Generate", "a", "vocabulary", "from", "txt", "files", "with", "example", "-", "per", "-", "line", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L450-L465
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
_shuffle_single
def _shuffle_single(fname, extra_fn=None): """Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling. """ records = read_records(fname) random.shuffle(records) if extra_fn is not None: ...
python
def _shuffle_single(fname, extra_fn=None): """Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling. """ records = read_records(fname) random.shuffle(records) if extra_fn is not None: ...
[ "def", "_shuffle_single", "(", "fname", ",", "extra_fn", "=", "None", ")", ":", "records", "=", "read_records", "(", "fname", ")", "random", ".", "shuffle", "(", "records", ")", "if", "extra_fn", "is", "not", "None", ":", "records", "=", "extra_fn", "(",...
Shuffle a single file of records. Args: fname: a string extra_fn: an optional function from list of TFRecords to list of TFRecords to be called after shuffling.
[ "Shuffle", "a", "single", "file", "of", "records", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L499-L513
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
shuffle_dataset
def shuffle_dataset(filenames, extra_fn=None): """Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file. """ if outputs_exist(filenames): tf.logging.info("Skipping shuffle because out...
python
def shuffle_dataset(filenames, extra_fn=None): """Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file. """ if outputs_exist(filenames): tf.logging.info("Skipping shuffle because out...
[ "def", "shuffle_dataset", "(", "filenames", ",", "extra_fn", "=", "None", ")", ":", "if", "outputs_exist", "(", "filenames", ")", ":", "tf", ".", "logging", ".", "info", "(", "\"Skipping shuffle because output files exist\"", ")", "return", "tf", ".", "logging",...
Shuffles the dataset. Args: filenames: a list of strings extra_fn: an optional function from list of records to list of records to be called after shuffling a file.
[ "Shuffles", "the", "dataset", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L516-L530
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
pack_examples
def pack_examples(examples, has_inputs, packed_length=256, spacing=2, queue_size=10, chop_long_sequences=False): """Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targe...
python
def pack_examples(examples, has_inputs, packed_length=256, spacing=2, queue_size=10, chop_long_sequences=False): """Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targe...
[ "def", "pack_examples", "(", "examples", ",", "has_inputs", ",", "packed_length", "=", "256", ",", "spacing", "=", "2", ",", "queue_size", "=", "10", ",", "chop_long_sequences", "=", "False", ")", ":", "packer", "=", "SequencePairPacker", "if", "has_inputs", ...
Pack examples into longer examples. If has_inputs=False, we are packing single-sequence examples with targets only and no inputs. In this case, we concatenate the targets from several examples to form each new example. We insert a number of zeros for spacing between the original sequences. This is to help...
[ "Pack", "examples", "into", "longer", "examples", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L589-L660
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
_pack_with_custom_ops
def _pack_with_custom_ops(dataset, keys, length): """Helper-function for packing a dataset which has already been batched. See pack_dataset() Relies on custom ops which require a custom compiled binary. Faster than _pack_with_tf_ops(), and denser packing. Args: dataset: a dataset containing padded batc...
python
def _pack_with_custom_ops(dataset, keys, length): """Helper-function for packing a dataset which has already been batched. See pack_dataset() Relies on custom ops which require a custom compiled binary. Faster than _pack_with_tf_ops(), and denser packing. Args: dataset: a dataset containing padded batc...
[ "def", "_pack_with_custom_ops", "(", "dataset", ",", "keys", ",", "length", ")", ":", "from", "tensor2tensor", ".", "data_generators", ".", "ops", "import", "pack_sequences_ops", "# pylint: disable=g-import-not-at-top", "# faster and better packing but requires custom-built bin...
Helper-function for packing a dataset which has already been batched. See pack_dataset() Relies on custom ops which require a custom compiled binary. Faster than _pack_with_tf_ops(), and denser packing. Args: dataset: a dataset containing padded batches of examples. keys: a list of strings (must have...
[ "Helper", "-", "function", "for", "packing", "a", "dataset", "which", "has", "already", "been", "batched", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L736-L770
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
make_tmp_dir
def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin """Make a temporary directory.""" if dir is None: return tempfile.mkdtemp(suffix, prefix, dir) else: while True: rand_term = random.randint(1, 9999) tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_te...
python
def make_tmp_dir(suffix="", prefix="tmp", dir=None): # pylint: disable=redefined-builtin """Make a temporary directory.""" if dir is None: return tempfile.mkdtemp(suffix, prefix, dir) else: while True: rand_term = random.randint(1, 9999) tmp_dir = os.path.join(dir, "%s%d%s" % (prefix, rand_te...
[ "def", "make_tmp_dir", "(", "suffix", "=", "\"\"", ",", "prefix", "=", "\"tmp\"", ",", "dir", "=", "None", ")", ":", "# pylint: disable=redefined-builtin", "if", "dir", "is", "None", ":", "return", "tempfile", ".", "mkdtemp", "(", "suffix", ",", "prefix", ...
Make a temporary directory.
[ "Make", "a", "temporary", "directory", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L883-L895
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
tfrecord_iterator_for_problem
def tfrecord_iterator_for_problem(problem, data_dir, dataset_split=tf.estimator.ModeKeys.TRAIN): """Iterate over the records on disk for the Problem.""" filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split)) example_spec = problem.example_reading_spec()[0] ...
python
def tfrecord_iterator_for_problem(problem, data_dir, dataset_split=tf.estimator.ModeKeys.TRAIN): """Iterate over the records on disk for the Problem.""" filenames = tf.gfile.Glob(problem.filepattern(data_dir, mode=dataset_split)) example_spec = problem.example_reading_spec()[0] ...
[ "def", "tfrecord_iterator_for_problem", "(", "problem", ",", "data_dir", ",", "dataset_split", "=", "tf", ".", "estimator", ".", "ModeKeys", ".", "TRAIN", ")", ":", "filenames", "=", "tf", ".", "gfile", ".", "Glob", "(", "problem", ".", "filepattern", "(", ...
Iterate over the records on disk for the Problem.
[ "Iterate", "over", "the", "records", "on", "disk", "for", "the", "Problem", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L898-L903
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
tfrecord_iterator
def tfrecord_iterator(filenames, gzipped=False, example_spec=None): """Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFe...
python
def tfrecord_iterator(filenames, gzipped=False, example_spec=None): """Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFe...
[ "def", "tfrecord_iterator", "(", "filenames", ",", "gzipped", "=", "False", ",", "example_spec", "=", "None", ")", ":", "with", "tf", ".", "Graph", "(", ")", ".", "as_default", "(", ")", ":", "dataset", "=", "tf", ".", "data", ".", "Dataset", ".", "f...
Yields records from TFRecord files. Args: filenames: list<str>, list of TFRecord filenames to read from. gzipped: bool, whether the TFRecord files are gzip-encoded. example_spec: dict<str feature name, tf.VarLenFeature/tf.FixedLenFeature>, if provided, will parse each record as a tensorflow.Example...
[ "Yields", "records", "from", "TFRecord", "files", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L906-L943
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/generator_utils.py
random_deinterleave
def random_deinterleave(text, separator_symbol="X"): """Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: t...
python
def random_deinterleave(text, separator_symbol="X"): """Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: t...
[ "def", "random_deinterleave", "(", "text", ",", "separator_symbol", "=", "\"X\"", ")", ":", "words", "=", "text", ".", "strip", "(", ")", ".", "split", "(", "\" \"", ")", "n", "=", "len", "(", "words", ")", "if", "n", "<=", "1", ":", "return", "tex...
Create a fill-in-the-blanks training example from text. Split on spaces, then cut into segments at random points. Alternate segments are assigned to the two output strings. separator_symbol separates segments within each of the outputs. example: text="The quick brown fox jumps over the lazy dog." ret...
[ "Create", "a", "fill", "-", "in", "-", "the", "-", "blanks", "training", "example", "from", "text", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/generator_utils.py#L946-L981
train
tensorflow/tensor2tensor
tensor2tensor/models/neural_gpu.py
neural_gpu_body
def neural_gpu_body(inputs, hparams, name=None): """The core Neural GPU.""" with tf.variable_scope(name, "neural_gpu"): def step(state, inp): # pylint: disable=missing-docstring x = tf.nn.dropout(state, 1.0 - hparams.dropout) for layer in range(hparams.num_hidden_layers): x = common_layers...
python
def neural_gpu_body(inputs, hparams, name=None): """The core Neural GPU.""" with tf.variable_scope(name, "neural_gpu"): def step(state, inp): # pylint: disable=missing-docstring x = tf.nn.dropout(state, 1.0 - hparams.dropout) for layer in range(hparams.num_hidden_layers): x = common_layers...
[ "def", "neural_gpu_body", "(", "inputs", ",", "hparams", ",", "name", "=", "None", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "\"neural_gpu\"", ")", ":", "def", "step", "(", "state", ",", "inp", ")", ":", "# pylint: disable=missing-do...
The core Neural GPU.
[ "The", "core", "Neural", "GPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/neural_gpu.py#L31-L52
train
tensorflow/tensor2tensor
tensor2tensor/models/neural_gpu.py
diagonal_neural_gpu
def diagonal_neural_gpu(inputs, hparams, name=None): """Improved Neural GPU as in https://arxiv.org/abs/1702.08727.""" with tf.variable_scope(name, "diagonal_neural_gpu"): def step(state_tup, inp): """Single step of the improved Neural GPU.""" state, _ = state_tup x = state for layer in...
python
def diagonal_neural_gpu(inputs, hparams, name=None): """Improved Neural GPU as in https://arxiv.org/abs/1702.08727.""" with tf.variable_scope(name, "diagonal_neural_gpu"): def step(state_tup, inp): """Single step of the improved Neural GPU.""" state, _ = state_tup x = state for layer in...
[ "def", "diagonal_neural_gpu", "(", "inputs", ",", "hparams", ",", "name", "=", "None", ")", ":", "with", "tf", ".", "variable_scope", "(", "name", ",", "\"diagonal_neural_gpu\"", ")", ":", "def", "step", "(", "state_tup", ",", "inp", ")", ":", "\"\"\"Singl...
Improved Neural GPU as in https://arxiv.org/abs/1702.08727.
[ "Improved", "Neural", "GPU", "as", "in", "https", ":", "//", "arxiv", ".", "org", "/", "abs", "/", "1702", ".", "08727", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/neural_gpu.py#L62-L87
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
_reorder_shape
def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name """Helper to determine the shape of reorder output.""" if output is None: return input_shape return base.nested_map(output, lambda i: input_shape[i])
python
def _reorder_shape(input_shape, output=None): # pylint: disable=invalid-name """Helper to determine the shape of reorder output.""" if output is None: return input_shape return base.nested_map(output, lambda i: input_shape[i])
[ "def", "_reorder_shape", "(", "input_shape", ",", "output", "=", "None", ")", ":", "# pylint: disable=invalid-name", "if", "output", "is", "None", ":", "return", "input_shape", "return", "base", ".", "nested_map", "(", "output", ",", "lambda", "i", ":", "input...
Helper to determine the shape of reorder output.
[ "Helper", "to", "determine", "the", "shape", "of", "reorder", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L77-L81
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
Reorder
def Reorder(x, params, output=None, **kwargs): """Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then ...
python
def Reorder(x, params, output=None, **kwargs): """Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then ...
[ "def", "Reorder", "(", "x", ",", "params", ",", "output", "=", "None", ",", "*", "*", "kwargs", ")", ":", "del", "params", ",", "kwargs", "if", "output", "is", "None", ":", "return", "x", "return", "base", ".", "nested_map", "(", "output", ",", "la...
Reorder a tuple into another tuple. For example, we can re-order (x, y) into (y, x) or even (y, (x, y), y). The output argument specifies how to re-order, using integers that refer to indices in the input tuple. For example, if input = (x, y, z) then Reorder(input, output=(1, 0, 2)) = (y, x, z) ...
[ "Reorder", "a", "tuple", "into", "another", "tuple", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L85-L115
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
_nested_op
def _nested_op(inputs, op): # pylint: disable=invalid-name """Helper: sum a list of arrays or nested arrays.""" # First the simple non-nested case. if not isinstance(inputs[0], (list, tuple)): return op(inputs) # In the nested case, sum on each axis separately. result_list = [] for i in range(len(input...
python
def _nested_op(inputs, op): # pylint: disable=invalid-name """Helper: sum a list of arrays or nested arrays.""" # First the simple non-nested case. if not isinstance(inputs[0], (list, tuple)): return op(inputs) # In the nested case, sum on each axis separately. result_list = [] for i in range(len(input...
[ "def", "_nested_op", "(", "inputs", ",", "op", ")", ":", "# pylint: disable=invalid-name", "# First the simple non-nested case.", "if", "not", "isinstance", "(", "inputs", "[", "0", "]", ",", "(", "list", ",", "tuple", ")", ")", ":", "return", "op", "(", "in...
Helper: sum a list of arrays or nested arrays.
[ "Helper", ":", "sum", "a", "list", "of", "arrays", "or", "nested", "arrays", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L134-L145
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
GateBranches
def GateBranches(x, **unused_kwargs): """Implements a gating function on a (memory, gate, candidate) tuple. Final update is memory * gate + (1-gate) * candidate This gating equation may also be referred to as Highway Network. Highway Networks: https://arxiv.org/abs/1505.00387 Args: x: A tuple of (memor...
python
def GateBranches(x, **unused_kwargs): """Implements a gating function on a (memory, gate, candidate) tuple. Final update is memory * gate + (1-gate) * candidate This gating equation may also be referred to as Highway Network. Highway Networks: https://arxiv.org/abs/1505.00387 Args: x: A tuple of (memor...
[ "def", "GateBranches", "(", "x", ",", "*", "*", "unused_kwargs", ")", ":", "assert", "len", "(", "x", ")", "==", "3", ",", "x", "state", ",", "gate", ",", "candidate", "=", "x", "return", "gate", "*", "state", "+", "(", "1.0", "-", "gate", ")", ...
Implements a gating function on a (memory, gate, candidate) tuple. Final update is memory * gate + (1-gate) * candidate This gating equation may also be referred to as Highway Network. Highway Networks: https://arxiv.org/abs/1505.00387 Args: x: A tuple of (memory, gate, candidate) Returns: The res...
[ "Implements", "a", "gating", "function", "on", "a", "(", "memory", "gate", "candidate", ")", "tuple", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L170-L186
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
_concatenate_shape
def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name """Helper to determine the shape of Concatenate output.""" ax = axis % len(input_shape[0]) concat_size = sum(shape[ax] for shape in input_shape) out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:] return out_sha...
python
def _concatenate_shape(input_shape, axis=-1): # pylint: disable=invalid-name """Helper to determine the shape of Concatenate output.""" ax = axis % len(input_shape[0]) concat_size = sum(shape[ax] for shape in input_shape) out_shape = input_shape[0][:ax] + (concat_size,) + input_shape[0][ax+1:] return out_sha...
[ "def", "_concatenate_shape", "(", "input_shape", ",", "axis", "=", "-", "1", ")", ":", "# pylint: disable=invalid-name", "ax", "=", "axis", "%", "len", "(", "input_shape", "[", "0", "]", ")", "concat_size", "=", "sum", "(", "shape", "[", "ax", "]", "for"...
Helper to determine the shape of Concatenate output.
[ "Helper", "to", "determine", "the", "shape", "of", "Concatenate", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L189-L194
train
tensorflow/tensor2tensor
tensor2tensor/trax/layers/combinators.py
Residual
def Residual(*layers, **kwargs): """Constructs a residual version of layers, summing input to layers output.""" shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter if len(layers) > 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Paralle...
python
def Residual(*layers, **kwargs): """Constructs a residual version of layers, summing input to layers output.""" shortcut = kwargs.get('shortcut', Identity()) # pylint: disable=no-value-for-parameter if len(layers) > 1: return Serial( Branch(), # pylint: disable=no-value-for-parameter Paralle...
[ "def", "Residual", "(", "*", "layers", ",", "*", "*", "kwargs", ")", ":", "shortcut", "=", "kwargs", ".", "get", "(", "'shortcut'", ",", "Identity", "(", ")", ")", "# pylint: disable=no-value-for-parameter", "if", "len", "(", "layers", ")", ">", "1", ":"...
Constructs a residual version of layers, summing input to layers output.
[ "Constructs", "a", "residual", "version", "of", "layers", "summing", "input", "to", "layers", "output", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/layers/combinators.py#L240-L256
train
tensorflow/tensor2tensor
tensor2tensor/rl/policy_learner.py
PolicyLearner.train
def train( self, env_fn, hparams, simulated, save_continuously, epoch, sampling_temp=1.0, num_env_steps=None, env_step_multiplier=1, eval_env_fn=None, report_fn=None ): """Train.""" raise NotImplementedError()
python
def train( self, env_fn, hparams, simulated, save_continuously, epoch, sampling_temp=1.0, num_env_steps=None, env_step_multiplier=1, eval_env_fn=None, report_fn=None ): """Train.""" raise NotImplementedError()
[ "def", "train", "(", "self", ",", "env_fn", ",", "hparams", ",", "simulated", ",", "save_continuously", ",", "epoch", ",", "sampling_temp", "=", "1.0", ",", "num_env_steps", "=", "None", ",", "env_step_multiplier", "=", "1", ",", "eval_env_fn", "=", "None", ...
Train.
[ "Train", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/rl/policy_learner.py#L34-L48
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
update_hparams_for_universal_transformer
def update_hparams_for_universal_transformer(hparams): """Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Univers...
python
def update_hparams_for_universal_transformer(hparams): """Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Univers...
[ "def", "update_hparams_for_universal_transformer", "(", "hparams", ")", ":", "hparams", ".", "daisy_chain_variables", "=", "False", "# Breaks multi-gpu in while loops.", "# If not None, mixes vanilla transformer with Universal Transformer.", "# Options: None, \"before_ut\", and \"after_ut\...
Adds default hparams for all of the variants of the Universal Transformer. Args: hparams: default hparams (usually one of the standard hparams from transformer model (like "transformer_base") Returns: hparams with default values for Universal Transformers hyper-parameters
[ "Adds", "default", "hparams", "for", "all", "of", "the", "variants", "of", "the", "Universal", "Transformer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L352-L436
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
universal_transformer_base
def universal_transformer_base(): """Base parameters for Universal Transformer.""" hparams = transformer.transformer_base() # To have a similar capacity to the transformer_base with 6 layers, # we need to increase the size of the UT's layer # since, in fact, UT has a single layer repeating multiple times. h...
python
def universal_transformer_base(): """Base parameters for Universal Transformer.""" hparams = transformer.transformer_base() # To have a similar capacity to the transformer_base with 6 layers, # we need to increase the size of the UT's layer # since, in fact, UT has a single layer repeating multiple times. h...
[ "def", "universal_transformer_base", "(", ")", ":", "hparams", "=", "transformer", ".", "transformer_base", "(", ")", "# To have a similar capacity to the transformer_base with 6 layers,", "# we need to increase the size of the UT's layer", "# since, in fact, UT has a single layer repeat...
Base parameters for Universal Transformer.
[ "Base", "parameters", "for", "Universal", "Transformer", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L440-L451
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
adaptive_universal_transformer_multilayer_tpu
def adaptive_universal_transformer_multilayer_tpu(): """Multi-layer config for adaptive Transformer on TPU.""" hparams = adaptive_universal_transformer_base_tpu() hparams.num_inrecurrence_layers = 2 hparams.mix_with_transformer = "before_ut,after_ut" hparams.num_mixedin_layers = 1 hparams.transformer_ffn_ty...
python
def adaptive_universal_transformer_multilayer_tpu(): """Multi-layer config for adaptive Transformer on TPU.""" hparams = adaptive_universal_transformer_base_tpu() hparams.num_inrecurrence_layers = 2 hparams.mix_with_transformer = "before_ut,after_ut" hparams.num_mixedin_layers = 1 hparams.transformer_ffn_ty...
[ "def", "adaptive_universal_transformer_multilayer_tpu", "(", ")", ":", "hparams", "=", "adaptive_universal_transformer_base_tpu", "(", ")", "hparams", ".", "num_inrecurrence_layers", "=", "2", "hparams", ".", "mix_with_transformer", "=", "\"before_ut,after_ut\"", "hparams", ...
Multi-layer config for adaptive Transformer on TPU.
[ "Multi", "-", "layer", "config", "for", "adaptive", "Transformer", "on", "TPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L543-L555
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
adaptive_universal_transformer_multilayer_hard
def adaptive_universal_transformer_multilayer_hard(): """Multi-layer config for adaptive Transformer with hard attention.""" hparams = adaptive_universal_transformer_multilayer_tpu() hparams.batch_size = 256 hparams.hard_attention_k = 8 hparams.add_step_timing_signal = True # hparams.add_sru = True # This ...
python
def adaptive_universal_transformer_multilayer_hard(): """Multi-layer config for adaptive Transformer with hard attention.""" hparams = adaptive_universal_transformer_multilayer_tpu() hparams.batch_size = 256 hparams.hard_attention_k = 8 hparams.add_step_timing_signal = True # hparams.add_sru = True # This ...
[ "def", "adaptive_universal_transformer_multilayer_hard", "(", ")", ":", "hparams", "=", "adaptive_universal_transformer_multilayer_tpu", "(", ")", "hparams", ".", "batch_size", "=", "256", "hparams", ".", "hard_attention_k", "=", "8", "hparams", ".", "add_step_timing_sign...
Multi-layer config for adaptive Transformer with hard attention.
[ "Multi", "-", "layer", "config", "for", "adaptive", "Transformer", "with", "hard", "attention", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L559-L568
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
universal_transformer_base_range
def universal_transformer_base_range(rhp): """Range of hyperparameters.""" # After starting from base, set intervals for some parameters. rhp.set_discrete("num_rec_steps", [6, 8, 10]) rhp.set_discrete("hidden_size", [1024, 2048, 4096]) rhp.set_discrete("filter_size", [2048, 4096, 8192]) rhp.set_discrete("nu...
python
def universal_transformer_base_range(rhp): """Range of hyperparameters.""" # After starting from base, set intervals for some parameters. rhp.set_discrete("num_rec_steps", [6, 8, 10]) rhp.set_discrete("hidden_size", [1024, 2048, 4096]) rhp.set_discrete("filter_size", [2048, 4096, 8192]) rhp.set_discrete("nu...
[ "def", "universal_transformer_base_range", "(", "rhp", ")", ":", "# After starting from base, set intervals for some parameters.", "rhp", ".", "set_discrete", "(", "\"num_rec_steps\"", ",", "[", "6", ",", "8", ",", "10", "]", ")", "rhp", ".", "set_discrete", "(", "\...
Range of hyperparameters.
[ "Range", "of", "hyperparameters", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L788-L797
train
tensorflow/tensor2tensor
tensor2tensor/models/research/universal_transformer.py
adaptive_universal_transformer_base_range
def adaptive_universal_transformer_base_range(rhp): """Range of hyperparameters.""" # After starting from base, set intervals for some parameters. rhp.set_discrete("act_max_steps", [8, 16, 32]) rhp.set_float("act_loss_weight", 0.0, 0.5) rhp.set_discrete("hidden_size", [1024, 2048, 4096]) rhp.set_discrete("f...
python
def adaptive_universal_transformer_base_range(rhp): """Range of hyperparameters.""" # After starting from base, set intervals for some parameters. rhp.set_discrete("act_max_steps", [8, 16, 32]) rhp.set_float("act_loss_weight", 0.0, 0.5) rhp.set_discrete("hidden_size", [1024, 2048, 4096]) rhp.set_discrete("f...
[ "def", "adaptive_universal_transformer_base_range", "(", "rhp", ")", ":", "# After starting from base, set intervals for some parameters.", "rhp", ".", "set_discrete", "(", "\"act_max_steps\"", ",", "[", "8", ",", "16", ",", "32", "]", ")", "rhp", ".", "set_float", "(...
Range of hyperparameters.
[ "Range", "of", "hyperparameters", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/models/research/universal_transformer.py#L801-L811
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/neural_gpu.py
DiagonalGate
def DiagonalGate(x, params, **kwargs): """Split channels in 3 parts. Shifts 1st and 3rd sections to left/right.""" del params del kwargs # x : [batch, 1, length, depth] x = np.pad( x, [(0, 0), (0, 0), (1, 1), (0, 0)], mode='constant', constant_values=0.0) depth = x.shape[-1] // 3 assert 3 * depth ==...
python
def DiagonalGate(x, params, **kwargs): """Split channels in 3 parts. Shifts 1st and 3rd sections to left/right.""" del params del kwargs # x : [batch, 1, length, depth] x = np.pad( x, [(0, 0), (0, 0), (1, 1), (0, 0)], mode='constant', constant_values=0.0) depth = x.shape[-1] // 3 assert 3 * depth ==...
[ "def", "DiagonalGate", "(", "x", ",", "params", ",", "*", "*", "kwargs", ")", ":", "del", "params", "del", "kwargs", "# x : [batch, 1, length, depth]", "x", "=", "np", ".", "pad", "(", "x", ",", "[", "(", "0", ",", "0", ")", ",", "(", "0", ",", "...
Split channels in 3 parts. Shifts 1st and 3rd sections to left/right.
[ "Split", "channels", "in", "3", "parts", ".", "Shifts", "1st", "and", "3rd", "sections", "to", "left", "/", "right", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L33-L47
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/neural_gpu.py
ConvDiagonalGRU
def ConvDiagonalGRU(units, kernel_size=(3, 3)): """Build convolutional GRU with diagonal gating as in ImprovedNGPU.""" def BuildConv(): return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME') return layers.GeneralGRUCell( candidate_transform=BuildConv, memory_transform=Diagona...
python
def ConvDiagonalGRU(units, kernel_size=(3, 3)): """Build convolutional GRU with diagonal gating as in ImprovedNGPU.""" def BuildConv(): return layers.Conv(filters=units, kernel_size=kernel_size, padding='SAME') return layers.GeneralGRUCell( candidate_transform=BuildConv, memory_transform=Diagona...
[ "def", "ConvDiagonalGRU", "(", "units", ",", "kernel_size", "=", "(", "3", ",", "3", ")", ")", ":", "def", "BuildConv", "(", ")", ":", "return", "layers", ".", "Conv", "(", "filters", "=", "units", ",", "kernel_size", "=", "kernel_size", ",", "padding"...
Build convolutional GRU with diagonal gating as in ImprovedNGPU.
[ "Build", "convolutional", "GRU", "with", "diagonal", "gating", "as", "in", "ImprovedNGPU", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L50-L60
train
tensorflow/tensor2tensor
tensor2tensor/trax/models/neural_gpu.py
NeuralGPU
def NeuralGPU(feature_depth=96, steps=16, vocab_size=2): """Implementation of Neural GPU: https://arxiv.org/abs/1702.08727. Args: feature_depth: Number of memory channels steps: Number of times depthwise recurrence steps. vocab_size: Vocabulary size. Returns: A NeuralGPU Stax model. """ xs =...
python
def NeuralGPU(feature_depth=96, steps=16, vocab_size=2): """Implementation of Neural GPU: https://arxiv.org/abs/1702.08727. Args: feature_depth: Number of memory channels steps: Number of times depthwise recurrence steps. vocab_size: Vocabulary size. Returns: A NeuralGPU Stax model. """ xs =...
[ "def", "NeuralGPU", "(", "feature_depth", "=", "96", ",", "steps", "=", "16", ",", "vocab_size", "=", "2", ")", ":", "xs", "=", "[", "]", "xs", ".", "append", "(", "layers", ".", "Embedding", "(", "feature_depth", "=", "feature_depth", ",", "vocab_size...
Implementation of Neural GPU: https://arxiv.org/abs/1702.08727. Args: feature_depth: Number of memory channels steps: Number of times depthwise recurrence steps. vocab_size: Vocabulary size. Returns: A NeuralGPU Stax model.
[ "Implementation", "of", "Neural", "GPU", ":", "https", ":", "//", "arxiv", ".", "org", "/", "abs", "/", "1702", ".", "08727", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/models/neural_gpu.py#L63-L82
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
strip_ids
def strip_ids(ids, ids_to_strip): """Strip ids_to_strip from the end ids.""" ids = list(ids) while ids and ids[-1] in ids_to_strip: ids.pop() return ids
python
def strip_ids(ids, ids_to_strip): """Strip ids_to_strip from the end ids.""" ids = list(ids) while ids and ids[-1] in ids_to_strip: ids.pop() return ids
[ "def", "strip_ids", "(", "ids", ",", "ids_to_strip", ")", ":", "ids", "=", "list", "(", "ids", ")", "while", "ids", "and", "ids", "[", "-", "1", "]", "in", "ids_to_strip", ":", "ids", ".", "pop", "(", ")", "return", "ids" ]
Strip ids_to_strip from the end ids.
[ "Strip", "ids_to_strip", "from", "the", "end", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L99-L104
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
_escape_token
def _escape_token(token, alphabet): """Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: ...
python
def _escape_token(token, alphabet): """Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: ...
[ "def", "_escape_token", "(", "token", ",", "alphabet", ")", ":", "if", "not", "isinstance", "(", "token", ",", "six", ".", "text_type", ")", ":", "raise", "ValueError", "(", "\"Expected string type for token, got %s\"", "%", "type", "(", "token", ")", ")", "...
Escape away underscores and OOV characters and append '_'. This allows the token to be expressed as the concatenation of a list of subtokens from the vocabulary. The underscore acts as a sentinel which allows us to invertibly concatenate multiple such lists. Args: token: A unicode string to be escaped. ...
[ "Escape", "away", "underscores", "and", "OOV", "characters", "and", "append", "_", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L400-L422
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.encode
def encode(self, s): """Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of ...
python
def encode(self, s): """Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of ...
[ "def", "encode", "(", "self", ",", "s", ")", ":", "return", "[", "int", "(", "w", ")", "+", "self", ".", "_num_reserved_ids", "for", "w", "in", "s", ".", "split", "(", ")", "]" ]
Transform a human-readable string into a sequence of int ids. The ids should be in the range [num_reserved_ids, vocab_size). Ids [0, num_reserved_ids) are reserved. EOS is not appended. Args: s: human-readable string to be converted. Returns: ids: list of integers
[ "Transform", "a", "human", "-", "readable", "string", "into", "a", "sequence", "of", "int", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L117-L131
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.decode
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: ...
python
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: ...
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "if", "strip_extraneous", ":", "ids", "=", "strip_ids", "(", "ids", ",", "list", "(", "range", "(", "self", ".", "_num_reserved_ids", "or", "0", ")", ")", ")", ...
Transform a sequence of int ids into a human-readable string. EOS is not expected in ids. Args: ids: list of integers to be converted. strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: s: human-readable string.
[ "Transform", "a", "sequence", "of", "int", "ids", "into", "a", "human", "-", "readable", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L133-L148
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TextEncoder.decode_list
def decode_list(self, ids): """Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integ...
python
def decode_list(self, ids): """Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integ...
[ "def", "decode_list", "(", "self", ",", "ids", ")", ":", "decoded_ids", "=", "[", "]", "for", "id_", "in", "ids", ":", "if", "0", "<=", "id_", "<", "self", ".", "_num_reserved_ids", ":", "decoded_ids", ".", "append", "(", "RESERVED_TOKENS", "[", "int",...
Transform a sequence of int ids into a their string versions. This method supports transforming individual input/output ids to their string versions so that sequence to/from text conversions can be visualized in a human readable format. Args: ids: list of integers to be converted. Returns: ...
[ "Transform", "a", "sequence", "of", "int", "ids", "into", "a", "their", "string", "versions", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L150-L169
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder.encode
def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] ...
python
def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] ...
[ "def", "encode", "(", "self", ",", "s", ")", ":", "sentence", "=", "s", "tokens", "=", "sentence", ".", "strip", "(", ")", ".", "split", "(", ")", "if", "self", ".", "_replace_oov", "is", "not", "None", ":", "tokens", "=", "[", "t", "if", "t", ...
Converts a space-separated string of tokens to a list of ids.
[ "Converts", "a", "space", "-", "separated", "string", "of", "tokens", "to", "a", "list", "of", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L314-L322
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder._init_vocab_from_file
def _init_vocab_from_file(self, filename): """Load vocab from a file. Args: filename: The file to load vocabulary from. """ with tf.gfile.Open(filename) as f: tokens = [token.strip() for token in f.readlines()] def token_gen(): for token in tokens: yield token self._...
python
def _init_vocab_from_file(self, filename): """Load vocab from a file. Args: filename: The file to load vocabulary from. """ with tf.gfile.Open(filename) as f: tokens = [token.strip() for token in f.readlines()] def token_gen(): for token in tokens: yield token self._...
[ "def", "_init_vocab_from_file", "(", "self", ",", "filename", ")", ":", "with", "tf", ".", "gfile", ".", "Open", "(", "filename", ")", "as", "f", ":", "tokens", "=", "[", "token", ".", "strip", "(", ")", "for", "token", "in", "f", ".", "readlines", ...
Load vocab from a file. Args: filename: The file to load vocabulary from.
[ "Load", "vocab", "from", "a", "file", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L338-L351
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder._init_vocab_from_list
def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token ...
python
def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token ...
[ "def", "_init_vocab_from_list", "(", "self", ",", "vocab_list", ")", ":", "def", "token_gen", "(", ")", ":", "for", "token", "in", "vocab_list", ":", "if", "token", "not", "in", "RESERVED_TOKENS", ":", "yield", "token", "self", ".", "_init_vocab", "(", "to...
Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens.
[ "Initialize", "tokens", "from", "a", "list", "of", "tokens", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L353-L367
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder._init_vocab
def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RE...
python
def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RE...
[ "def", "_init_vocab", "(", "self", ",", "token_generator", ",", "add_reserved_tokens", "=", "True", ")", ":", "self", ".", "_id_to_token", "=", "{", "}", "non_reserved_start_index", "=", "0", "if", "add_reserved_tokens", ":", "self", ".", "_id_to_token", ".", ...
Initialize vocabulary with tokens from token_generator.
[ "Initialize", "vocabulary", "with", "tokens", "from", "token_generator", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L369-L384
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
TokenTextEncoder.store_to_file
def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with tf.gfile.Open(filename, "w") as f:...
python
def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with tf.gfile.Open(filename, "w") as f:...
[ "def", "store_to_file", "(", "self", ",", "filename", ")", ":", "with", "tf", ".", "gfile", ".", "Open", "(", "filename", ",", "\"w\"", ")", "as", "f", ":", "for", "i", "in", "range", "(", "len", "(", "self", ".", "_id_to_token", ")", ")", ":", "...
Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to.
[ "Write", "vocab", "file", "to", "disk", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L386-L397
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.decode
def decode(self, ids, strip_extraneous=False): """Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string """ ...
python
def decode(self, ids, strip_extraneous=False): """Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string """ ...
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "if", "strip_extraneous", ":", "ids", "=", "strip_ids", "(", "ids", ",", "list", "(", "range", "(", "self", ".", "_num_reserved_ids", "or", "0", ")", ")", ")", ...
Converts a sequence of subtoken ids to a native string. Args: ids: a list of integers in the range [0, vocab_size) strip_extraneous: bool, whether to strip off extraneous tokens (EOS and PAD). Returns: a native string
[ "Converts", "a", "sequence", "of", "subtoken", "ids", "to", "a", "native", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L522-L536
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._tokens_to_subtoken_ids
def _tokens_to_subtoken_ids(self, tokens): """Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size) """ ret = [] for token in tokens: ret.extend(self._token_to_subtoken_ids(token)) ret...
python
def _tokens_to_subtoken_ids(self, tokens): """Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size) """ ret = [] for token in tokens: ret.extend(self._token_to_subtoken_ids(token)) ret...
[ "def", "_tokens_to_subtoken_ids", "(", "self", ",", "tokens", ")", ":", "ret", "=", "[", "]", "for", "token", "in", "tokens", ":", "ret", ".", "extend", "(", "self", ".", "_token_to_subtoken_ids", "(", "token", ")", ")", "return", "ret" ]
Converts a list of tokens to a list of subtoken ids. Args: tokens: a list of strings. Returns: a list of integers in the range [0, vocab_size)
[ "Converts", "a", "list", "of", "tokens", "to", "a", "list", "of", "subtoken", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L546-L557
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._token_to_subtoken_ids
def _token_to_subtoken_ids(self, token): """Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size) """ cache_location = hash(token) % self._cache_size cache_key, cache_value = self._cache[cache_location] if cache...
python
def _token_to_subtoken_ids(self, token): """Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size) """ cache_location = hash(token) % self._cache_size cache_key, cache_value = self._cache[cache_location] if cache...
[ "def", "_token_to_subtoken_ids", "(", "self", ",", "token", ")", ":", "cache_location", "=", "hash", "(", "token", ")", "%", "self", ".", "_cache_size", "cache_key", ",", "cache_value", "=", "self", ".", "_cache", "[", "cache_location", "]", "if", "cache_key...
Converts token to a list of subtoken ids. Args: token: a string. Returns: a list of integers in the range [0, vocab_size)
[ "Converts", "token", "to", "a", "list", "of", "subtoken", "ids", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L559-L574
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._subtoken_ids_to_tokens
def _subtoken_ids_to_tokens(self, subtokens): """Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings. """ concatenated = "".join( [self._subtoken_id_to_subtoken_string(s) for s in subtok...
python
def _subtoken_ids_to_tokens(self, subtokens): """Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings. """ concatenated = "".join( [self._subtoken_id_to_subtoken_string(s) for s in subtok...
[ "def", "_subtoken_ids_to_tokens", "(", "self", ",", "subtokens", ")", ":", "concatenated", "=", "\"\"", ".", "join", "(", "[", "self", ".", "_subtoken_id_to_subtoken_string", "(", "s", ")", "for", "s", "in", "subtokens", "]", ")", "split", "=", "concatenated...
Converts a list of subtoken ids to a list of tokens. Args: subtokens: a list of integers in the range [0, vocab_size) Returns: a list of strings.
[ "Converts", "a", "list", "of", "subtoken", "ids", "to", "a", "list", "of", "tokens", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L576-L593
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._subtoken_id_to_subtoken_string
def _subtoken_id_to_subtoken_string(self, subtoken): """Converts a subtoken integer ID to a subtoken string.""" if 0 <= subtoken < self.vocab_size: return self._all_subtoken_strings[subtoken] return u""
python
def _subtoken_id_to_subtoken_string(self, subtoken): """Converts a subtoken integer ID to a subtoken string.""" if 0 <= subtoken < self.vocab_size: return self._all_subtoken_strings[subtoken] return u""
[ "def", "_subtoken_id_to_subtoken_string", "(", "self", ",", "subtoken", ")", ":", "if", "0", "<=", "subtoken", "<", "self", ".", "vocab_size", ":", "return", "self", ".", "_all_subtoken_strings", "[", "subtoken", "]", "return", "u\"\"" ]
Converts a subtoken integer ID to a subtoken string.
[ "Converts", "a", "subtoken", "integer", "ID", "to", "a", "subtoken", "string", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L595-L599
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._escaped_token_to_subtoken_strings
def _escaped_token_to_subtoken_strings(self, escaped_token): """Converts an escaped token string to a list of subtoken strings. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtokens as unicode strings. """ # NOTE: This algorithm is greedy; it won't nece...
python
def _escaped_token_to_subtoken_strings(self, escaped_token): """Converts an escaped token string to a list of subtoken strings. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtokens as unicode strings. """ # NOTE: This algorithm is greedy; it won't nece...
[ "def", "_escaped_token_to_subtoken_strings", "(", "self", ",", "escaped_token", ")", ":", "# NOTE: This algorithm is greedy; it won't necessarily produce the \"best\"", "# list of subtokens.", "ret", "=", "[", "]", "start", "=", "0", "token_len", "=", "len", "(", "escaped_t...
Converts an escaped token string to a list of subtoken strings. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtokens as unicode strings.
[ "Converts", "an", "escaped", "token", "string", "to", "a", "list", "of", "subtoken", "strings", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L601-L629
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._escaped_token_to_subtoken_ids
def _escaped_token_to_subtoken_ids(self, escaped_token): """Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers. """ return [ self._subtoken_string_to_id[subtoken] ...
python
def _escaped_token_to_subtoken_ids(self, escaped_token): """Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers. """ return [ self._subtoken_string_to_id[subtoken] ...
[ "def", "_escaped_token_to_subtoken_ids", "(", "self", ",", "escaped_token", ")", ":", "return", "[", "self", ".", "_subtoken_string_to_id", "[", "subtoken", "]", "for", "subtoken", "in", "self", ".", "_escaped_token_to_subtoken_strings", "(", "escaped_token", ")", "...
Converts an escaped token string to a list of subtoken IDs. Args: escaped_token: An escaped token as a unicode string. Returns: A list of subtoken IDs as integers.
[ "Converts", "an", "escaped", "token", "string", "to", "a", "list", "of", "subtoken", "IDs", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L631-L642
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_from_generator
def build_from_generator(cls, generator, target_size, max_subtoken_length=None, reserved_tokens=None): """Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. ta...
python
def build_from_generator(cls, generator, target_size, max_subtoken_length=None, reserved_tokens=None): """Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. ta...
[ "def", "build_from_generator", "(", "cls", ",", "generator", ",", "target_size", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ")", ":", "token_counts", "=", "collections", ".", "defaultdict", "(", "int", ")", "for", "item", "in...
Builds a SubwordTextEncoder from the generated text. Args: generator: yields text. target_size: int, approximate vocabulary size to create. max_subtoken_length: Maximum length of a subtoken. If this is not set, then the runtime and memory use of creating the vocab is quadratic in ...
[ "Builds", "a", "SubwordTextEncoder", "from", "the", "generated", "text", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L645-L674
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_to_target_size
def build_to_target_size(cls, target_size, token_counts, min_val, max_val, max_subtoken_length=None, reserved_tokens=None, num_iter...
python
def build_to_target_size(cls, target_size, token_counts, min_val, max_val, max_subtoken_length=None, reserved_tokens=None, num_iter...
[ "def", "build_to_target_size", "(", "cls", ",", "target_size", ",", "token_counts", ",", "min_val", ",", "max_val", ",", "max_subtoken_length", "=", "None", ",", "reserved_tokens", "=", "None", ",", "num_iterations", "=", "4", ")", ":", "if", "min_val", ">", ...
Builds a SubwordTextEncoder that has `vocab_size` near `target_size`. Uses simple recursive binary search to find a minimum token count that most closely matches the `target_size`. Args: target_size: Desired vocab_size to approximate. token_counts: A dictionary of token counts, mapping string ...
[ "Builds", "a", "SubwordTextEncoder", "that", "has", "vocab_size", "near", "target_size", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L677-L748
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.build_from_token_counts
def build_from_token_counts(self, token_counts, min_count, num_iterations=4, reserved_tokens=None, max_subtoken_length=None): """Train a SubwordTextEncoder based on a...
python
def build_from_token_counts(self, token_counts, min_count, num_iterations=4, reserved_tokens=None, max_subtoken_length=None): """Train a SubwordTextEncoder based on a...
[ "def", "build_from_token_counts", "(", "self", ",", "token_counts", ",", "min_count", ",", "num_iterations", "=", "4", ",", "reserved_tokens", "=", "None", ",", "max_subtoken_length", "=", "None", ")", ":", "if", "reserved_tokens", "is", "None", ":", "reserved_t...
Train a SubwordTextEncoder based on a dictionary of word counts. Args: token_counts: a dictionary of Unicode strings to int. min_count: an integer - discard subtokens with lower counts. num_iterations: an integer. how many iterations of refinement. reserved_tokens: List of reserved tokens....
[ "Train", "a", "SubwordTextEncoder", "based", "on", "a", "dictionary", "of", "word", "counts", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L750-L866
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder.dump
def dump(self): """Debugging dump of the current subtoken vocabulary.""" subtoken_strings = [(i, s) for s, i in six.iteritems(self._subtoken_string_to_id)] print(u", ".join(u"{0} : '{1}'".format(i, s) for i, s in sorted(subtoken_strings)))
python
def dump(self): """Debugging dump of the current subtoken vocabulary.""" subtoken_strings = [(i, s) for s, i in six.iteritems(self._subtoken_string_to_id)] print(u", ".join(u"{0} : '{1}'".format(i, s) for i, s in sorted(subtoken_strings)))
[ "def", "dump", "(", "self", ")", ":", "subtoken_strings", "=", "[", "(", "i", ",", "s", ")", "for", "s", ",", "i", "in", "six", ".", "iteritems", "(", "self", ".", "_subtoken_string_to_id", ")", "]", "print", "(", "u\", \"", ".", "join", "(", "u\"{...
Debugging dump of the current subtoken vocabulary.
[ "Debugging", "dump", "of", "the", "current", "subtoken", "vocabulary", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L872-L877
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._init_subtokens_from_list
def _init_subtokens_from_list(self, subtoken_strings, reserved_tokens=None): """Initialize token information from a list of subtoken strings. Args: subtoken_strings: a list of subtokens reserved_tokens: List of reserved tokens. We must have `reserved_tokens` as None or the empty list, or el...
python
def _init_subtokens_from_list(self, subtoken_strings, reserved_tokens=None): """Initialize token information from a list of subtoken strings. Args: subtoken_strings: a list of subtokens reserved_tokens: List of reserved tokens. We must have `reserved_tokens` as None or the empty list, or el...
[ "def", "_init_subtokens_from_list", "(", "self", ",", "subtoken_strings", ",", "reserved_tokens", "=", "None", ")", ":", "if", "reserved_tokens", "is", "None", ":", "reserved_tokens", "=", "[", "]", "if", "reserved_tokens", ":", "self", ".", "_all_subtoken_strings...
Initialize token information from a list of subtoken strings. Args: subtoken_strings: a list of subtokens reserved_tokens: List of reserved tokens. We must have `reserved_tokens` as None or the empty list, or else the global variable `RESERVED_TOKENS` must be a prefix of `reserved_token...
[ "Initialize", "token", "information", "from", "a", "list", "of", "subtoken", "strings", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L879-L910
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._load_from_file_object
def _load_from_file_object(self, f): """Load from a file object. Args: f: File object to load vocabulary from """ subtoken_strings = [] for line in f: s = line.strip() # Some vocab files wrap words in single quotes, but others don't if ((s.startswith("'") and s.endswith("'")...
python
def _load_from_file_object(self, f): """Load from a file object. Args: f: File object to load vocabulary from """ subtoken_strings = [] for line in f: s = line.strip() # Some vocab files wrap words in single quotes, but others don't if ((s.startswith("'") and s.endswith("'")...
[ "def", "_load_from_file_object", "(", "self", ",", "f", ")", ":", "subtoken_strings", "=", "[", "]", "for", "line", "in", "f", ":", "s", "=", "line", ".", "strip", "(", ")", "# Some vocab files wrap words in single quotes, but others don't", "if", "(", "(", "s...
Load from a file object. Args: f: File object to load vocabulary from
[ "Load", "from", "a", "file", "object", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L919-L934
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
SubwordTextEncoder._load_from_file
def _load_from_file(self, filename): """Load from a vocab file.""" if not tf.gfile.Exists(filename): raise ValueError("File %s not found" % filename) with tf.gfile.Open(filename) as f: self._load_from_file_object(f)
python
def _load_from_file(self, filename): """Load from a vocab file.""" if not tf.gfile.Exists(filename): raise ValueError("File %s not found" % filename) with tf.gfile.Open(filename) as f: self._load_from_file_object(f)
[ "def", "_load_from_file", "(", "self", ",", "filename", ")", ":", "if", "not", "tf", ".", "gfile", ".", "Exists", "(", "filename", ")", ":", "raise", "ValueError", "(", "\"File %s not found\"", "%", "filename", ")", "with", "tf", ".", "gfile", ".", "Open...
Load from a vocab file.
[ "Load", "from", "a", "vocab", "file", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L936-L941
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
ImageEncoder.encode
def encode(self, s): """Transform a string with a filename into a list of RGB integers. Args: s: path to the file with an image. Returns: ids: list of integers """ try: import matplotlib.image as im # pylint: disable=g-import-not-at-top except ImportError as e: tf.logg...
python
def encode(self, s): """Transform a string with a filename into a list of RGB integers. Args: s: path to the file with an image. Returns: ids: list of integers """ try: import matplotlib.image as im # pylint: disable=g-import-not-at-top except ImportError as e: tf.logg...
[ "def", "encode", "(", "self", ",", "s", ")", ":", "try", ":", "import", "matplotlib", ".", "image", "as", "im", "# pylint: disable=g-import-not-at-top", "except", "ImportError", "as", "e", ":", "tf", ".", "logging", ".", "warning", "(", "\"Reading an image req...
Transform a string with a filename into a list of RGB integers. Args: s: path to the file with an image. Returns: ids: list of integers
[ "Transform", "a", "string", "with", "a", "filename", "into", "a", "list", "of", "RGB", "integers", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L965-L980
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
ImageEncoder.decode
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into an image file. Args: ids: list of integers to be converted. strip_extraneous: unused Returns: Path to the temporary file where the image was saved. Raises: ValueError: if the ids are not of ...
python
def decode(self, ids, strip_extraneous=False): """Transform a sequence of int ids into an image file. Args: ids: list of integers to be converted. strip_extraneous: unused Returns: Path to the temporary file where the image was saved. Raises: ValueError: if the ids are not of ...
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "del", "strip_extraneous", "_", ",", "tmp_file_path", "=", "tempfile", ".", "mkstemp", "(", "\"_decode.png\"", ")", "if", "self", ".", "_height", "is", "None", "or",...
Transform a sequence of int ids into an image file. Args: ids: list of integers to be converted. strip_extraneous: unused Returns: Path to the temporary file where the image was saved. Raises: ValueError: if the ids are not of the appropriate size.
[ "Transform", "a", "sequence", "of", "int", "ids", "into", "an", "image", "file", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L982-L1018
train
tensorflow/tensor2tensor
tensor2tensor/data_generators/text_encoder.py
RealEncoder.decode
def decode(self, ids, strip_extraneous=False): """Transform sequence of float values into string (float values). Args: ids: array of floats to be converted. strip_extraneous: unused Returns: String having space separated float values. Raises: ValueError: if the ids are not of ...
python
def decode(self, ids, strip_extraneous=False): """Transform sequence of float values into string (float values). Args: ids: array of floats to be converted. strip_extraneous: unused Returns: String having space separated float values. Raises: ValueError: if the ids are not of ...
[ "def", "decode", "(", "self", ",", "ids", ",", "strip_extraneous", "=", "False", ")", ":", "del", "strip_extraneous", "return", "\" \"", ".", "join", "(", "[", "str", "(", "i", ")", "for", "i", "in", "ids", "]", ")" ]
Transform sequence of float values into string (float values). Args: ids: array of floats to be converted. strip_extraneous: unused Returns: String having space separated float values. Raises: ValueError: if the ids are not of the appropriate size.
[ "Transform", "sequence", "of", "float", "values", "into", "string", "(", "float", "values", ")", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/data_generators/text_encoder.py#L1050-L1064
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
_pack_images
def _pack_images(images, rows, cols): """Helper utility to make a tiled field of images from numpy arrays. Args: images: Image tensor in shape [N, W, H, C]. rows: Number of images per row in tiled image. cols: Number of images per column in tiled image. Returns: A tiled image of shape [W * rows,...
python
def _pack_images(images, rows, cols): """Helper utility to make a tiled field of images from numpy arrays. Args: images: Image tensor in shape [N, W, H, C]. rows: Number of images per row in tiled image. cols: Number of images per column in tiled image. Returns: A tiled image of shape [W * rows,...
[ "def", "_pack_images", "(", "images", ",", "rows", ",", "cols", ")", ":", "shape", "=", "onp", ".", "shape", "(", "images", ")", "width", ",", "height", ",", "depth", "=", "shape", "[", "-", "3", ":", "]", "images", "=", "onp", ".", "reshape", "(...
Helper utility to make a tiled field of images from numpy arrays. Args: images: Image tensor in shape [N, W, H, C]. rows: Number of images per row in tiled image. cols: Number of images per column in tiled image. Returns: A tiled image of shape [W * rows, H * cols, C]. Truncates incomplete row...
[ "Helper", "utility", "to", "make", "a", "tiled", "field", "of", "images", "from", "numpy", "arrays", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L49-L71
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
markdownify_operative_config_str
def markdownify_operative_config_str(string): """Convert an operative config string to markdown format.""" # TODO(b/37527917): Total hack below. Implement more principled formatting. def process(line): """Convert a single line to markdown format.""" if not line.startswith('#'): return ' ' + line...
python
def markdownify_operative_config_str(string): """Convert an operative config string to markdown format.""" # TODO(b/37527917): Total hack below. Implement more principled formatting. def process(line): """Convert a single line to markdown format.""" if not line.startswith('#'): return ' ' + line...
[ "def", "markdownify_operative_config_str", "(", "string", ")", ":", "# TODO(b/37527917): Total hack below. Implement more principled formatting.", "def", "process", "(", "line", ")", ":", "\"\"\"Convert a single line to markdown format.\"\"\"", "if", "not", "line", ".", "startswi...
Convert an operative config string to markdown format.
[ "Convert", "an", "operative", "config", "string", "to", "markdown", "format", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L326-L350
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.close
def close(self): """Close SummaryWriter. Final!""" if not self._closed: self._event_writer.close() self._closed = True del self._event_writer
python
def close(self): """Close SummaryWriter. Final!""" if not self._closed: self._event_writer.close() self._closed = True del self._event_writer
[ "def", "close", "(", "self", ")", ":", "if", "not", "self", ".", "_closed", ":", "self", ".", "_event_writer", ".", "close", "(", ")", "self", ".", "_closed", "=", "True", "del", "self", ".", "_event_writer" ]
Close SummaryWriter. Final!
[ "Close", "SummaryWriter", ".", "Final!" ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L98-L103
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.scalar
def scalar(self, tag, value, step=None): """Saves scalar value. Args: tag: str: label for this data value: int/float: number to log step: int: training step """ value = float(onp.array(value)) if step is None: step = self._step else: self._step = step summary =...
python
def scalar(self, tag, value, step=None): """Saves scalar value. Args: tag: str: label for this data value: int/float: number to log step: int: training step """ value = float(onp.array(value)) if step is None: step = self._step else: self._step = step summary =...
[ "def", "scalar", "(", "self", ",", "tag", ",", "value", ",", "step", "=", "None", ")", ":", "value", "=", "float", "(", "onp", ".", "array", "(", "value", ")", ")", "if", "step", "is", "None", ":", "step", "=", "self", ".", "_step", "else", ":"...
Saves scalar value. Args: tag: str: label for this data value: int/float: number to log step: int: training step
[ "Saves", "scalar", "value", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L111-L125
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.image
def image(self, tag, image, step=None): """Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3]. Args: tag: str: label for this data image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/ step: int: training step """ image = onp.array(image) if s...
python
def image(self, tag, image, step=None): """Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3]. Args: tag: str: label for this data image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/ step: int: training step """ image = onp.array(image) if s...
[ "def", "image", "(", "self", ",", "tag", ",", "image", ",", "step", "=", "None", ")", ":", "image", "=", "onp", ".", "array", "(", "image", ")", "if", "step", "is", "None", ":", "step", "=", "self", ".", "_step", "else", ":", "self", ".", "_ste...
Saves RGB image summary from onp.ndarray [H,W], [H,W,1], or [H,W,3]. Args: tag: str: label for this data image: ndarray: [H,W], [H,W,1], [H,W,3] save image in greyscale or colors/ step: int: training step
[ "Saves", "RGB", "image", "summary", "from", "onp", ".", "ndarray", "[", "H", "W", "]", "[", "H", "W", "1", "]", "or", "[", "H", "W", "3", "]", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L127-L152
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.images
def images(self, tag, images, step=None, rows=None, cols=None): """Saves (rows, cols) tiled images from onp.ndarray. If either rows or cols aren't given, they are determined automatically from the size of the image batch, if neither are given a long column of images is produced. This truncates the imag...
python
def images(self, tag, images, step=None, rows=None, cols=None): """Saves (rows, cols) tiled images from onp.ndarray. If either rows or cols aren't given, they are determined automatically from the size of the image batch, if neither are given a long column of images is produced. This truncates the imag...
[ "def", "images", "(", "self", ",", "tag", ",", "images", ",", "step", "=", "None", ",", "rows", "=", "None", ",", "cols", "=", "None", ")", ":", "images", "=", "onp", ".", "array", "(", "images", ")", "if", "step", "is", "None", ":", "step", "=...
Saves (rows, cols) tiled images from onp.ndarray. If either rows or cols aren't given, they are determined automatically from the size of the image batch, if neither are given a long column of images is produced. This truncates the image batch rather than padding if it doesn't fill the final row. ...
[ "Saves", "(", "rows", "cols", ")", "tiled", "images", "from", "onp", ".", "ndarray", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L154-L183
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.plot
def plot(self, tag, mpl_plt, step=None, close_plot=True): """Saves matplotlib plot output to summary image. Args: tag: str: label for this data mpl_plt: matplotlib stateful pyplot object with prepared plotting state step: int: training step close_plot: bool: automatically closes plot ...
python
def plot(self, tag, mpl_plt, step=None, close_plot=True): """Saves matplotlib plot output to summary image. Args: tag: str: label for this data mpl_plt: matplotlib stateful pyplot object with prepared plotting state step: int: training step close_plot: bool: automatically closes plot ...
[ "def", "plot", "(", "self", ",", "tag", ",", "mpl_plt", ",", "step", "=", "None", ",", "close_plot", "=", "True", ")", ":", "if", "step", "is", "None", ":", "step", "=", "self", ".", "_step", "else", ":", "self", ".", "_step", "=", "step", "fig",...
Saves matplotlib plot output to summary image. Args: tag: str: label for this data mpl_plt: matplotlib stateful pyplot object with prepared plotting state step: int: training step close_plot: bool: automatically closes plot
[ "Saves", "matplotlib", "plot", "output", "to", "summary", "image", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L185-L210
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.audio
def audio(self, tag, audiodata, step=None, sample_rate=44100): """Saves audio. NB: single channel only right now. Args: tag: str: label for this data audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave step: int: training step sample_rate: sample rate of passed ...
python
def audio(self, tag, audiodata, step=None, sample_rate=44100): """Saves audio. NB: single channel only right now. Args: tag: str: label for this data audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave step: int: training step sample_rate: sample rate of passed ...
[ "def", "audio", "(", "self", ",", "tag", ",", "audiodata", ",", "step", "=", "None", ",", "sample_rate", "=", "44100", ")", ":", "audiodata", "=", "onp", ".", "array", "(", "audiodata", ")", "if", "step", "is", "None", ":", "step", "=", "self", "."...
Saves audio. NB: single channel only right now. Args: tag: str: label for this data audiodata: ndarray [Nsamples,]: data between (-1.0,1.0) to save as wave step: int: training step sample_rate: sample rate of passed in audio buffer
[ "Saves", "audio", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L212-L249
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.histogram
def histogram(self, tag, values, bins, step=None): """Saves histogram of values. Args: tag: str: label for this data values: ndarray: will be flattened by this routine bins: number of bins in histogram, or array of bins for onp.histogram step: int: training step """ if step is N...
python
def histogram(self, tag, values, bins, step=None): """Saves histogram of values. Args: tag: str: label for this data values: ndarray: will be flattened by this routine bins: number of bins in histogram, or array of bins for onp.histogram step: int: training step """ if step is N...
[ "def", "histogram", "(", "self", ",", "tag", ",", "values", ",", "bins", ",", "step", "=", "None", ")", ":", "if", "step", "is", "None", ":", "step", "=", "self", ".", "_step", "else", ":", "self", ".", "_step", "=", "step", "values", "=", "onp",...
Saves histogram of values. Args: tag: str: label for this data values: ndarray: will be flattened by this routine bins: number of bins in histogram, or array of bins for onp.histogram step: int: training step
[ "Saves", "histogram", "of", "values", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L251-L287
train
tensorflow/tensor2tensor
tensor2tensor/trax/jaxboard.py
SummaryWriter.text
def text(self, tag, textdata, step=None): """Saves a text summary. Args: tag: str: label for this data textdata: string, or 1D/2D list/numpy array of strings step: int: training step Note: markdown formatting is rendered by tensorboard. """ if step is None: step = self._step...
python
def text(self, tag, textdata, step=None): """Saves a text summary. Args: tag: str: label for this data textdata: string, or 1D/2D list/numpy array of strings step: int: training step Note: markdown formatting is rendered by tensorboard. """ if step is None: step = self._step...
[ "def", "text", "(", "self", ",", "tag", ",", "textdata", ",", "step", "=", "None", ")", ":", "if", "step", "is", "None", ":", "step", "=", "self", ".", "_step", "else", ":", "self", ".", "_step", "=", "step", "smd", "=", "SummaryMetadata", "(", "...
Saves a text summary. Args: tag: str: label for this data textdata: string, or 1D/2D list/numpy array of strings step: int: training step Note: markdown formatting is rendered by tensorboard.
[ "Saves", "a", "text", "summary", "." ]
272500b6efe353aeb638d2745ed56e519462ca31
https://github.com/tensorflow/tensor2tensor/blob/272500b6efe353aeb638d2745ed56e519462ca31/tensor2tensor/trax/jaxboard.py#L289-L322
train