| import tensorflow as tf
|
| from .encoder import Encoder
|
| from .decoder import Decoder
|
| from tensorflow.keras.layers import Dense
|
|
|
| @tf.keras.utils.register_keras_serializable()
|
| class Transformer(tf.keras.Model):
|
| def __init__(self, num_layers, d_model, num_heads, dff, input_vocab_size,
|
| target_vocab_size, max_tokens, dropout_rate=0.1, **kwargs):
|
| super(Transformer, self).__init__(**kwargs)
|
| self.num_layers = num_layers
|
| self.d_model = d_model
|
| self.num_heads = num_heads
|
| self.dff = dff
|
| self.input_vocab_size = input_vocab_size
|
| self.target_vocab_size = target_vocab_size
|
| self.max_tokens = max_tokens
|
| self.dropout_rate = dropout_rate
|
|
|
| self.encoder = Encoder(num_layers, d_model, num_heads, dff,
|
| input_vocab_size, max_tokens, dropout_rate)
|
| self.decoder = Decoder(num_layers, d_model, num_heads, dff,
|
| target_vocab_size, max_tokens, dropout_rate)
|
| self.final_layer = Dense(target_vocab_size)
|
|
|
| def call(self, inputs, training=None):
|
| enc_input, dec_input = inputs
|
| enc_padding_mask = self.create_padding_mask(enc_input)
|
| look_ahead_mask = self.create_look_ahead_mask(tf.shape(dec_input)[1])
|
| dec_padding_mask = self.create_padding_mask(enc_input)
|
| enc_output = self.encoder(enc_input, training=training, mask=enc_padding_mask)
|
| dec_output = self.decoder(dec_input, enc_output, training=training,
|
| look_ahead_mask=look_ahead_mask,
|
| padding_mask=dec_padding_mask)
|
| final_output = self.final_layer(dec_output)
|
| return final_output
|
|
|
| def create_padding_mask(self, seq):
|
| mask = tf.cast(tf.math.equal(seq, 0), tf.float32)
|
| return mask[:, tf.newaxis, tf.newaxis, :]
|
|
|
| def create_look_ahead_mask(self, size):
|
| mask = 1 - tf.linalg.band_part(tf.ones((size, size)), -1, 0)
|
| return mask
|
|
|
| def get_config(self):
|
| config = super().get_config()
|
| config.update({
|
| 'num_layers': self.num_layers,
|
| 'd_model': self.d_model,
|
| 'num_heads': self.num_heads,
|
| 'dff': self.dff,
|
| 'input_vocab_size': self.input_vocab_size,
|
| 'target_vocab_size': self.target_vocab_size,
|
| 'max_tokens': self.max_tokens,
|
| 'dropout_rate': self.dropout_rate
|
| })
|
| return config |