Yuchan commited on
Commit
1f4c0fc
·
verified ·
1 Parent(s): cefdbfc

Update AlphaS2S.py

Browse files
Files changed (1) hide show
  1. AlphaS2S.py +2 -2
AlphaS2S.py CHANGED
@@ -196,7 +196,7 @@ class EncoderBlock(layers.Layer):
196
  def __init__(self, d_model, num_heads, dff, dropout=0.1):
197
  super().__init__()
198
  self.mha = layers.MultiHeadAttention(num_heads=num_heads, key_dim=d_model)
199
- self.ffn = SwiGLU(d_model, 512)
200
  self.norm1 = layers.LayerNormalization(epsilon=1e-6)
201
  self.norm2 = layers.LayerNormalization(epsilon=1e-6)
202
  self.dropout1 = layers.Dropout(dropout)
@@ -222,7 +222,7 @@ class LoU(layers.Layer):
222
  self.alpha_linear = layers.Dense(1, activation='sigmoid', dtype='float32')
223
 
224
  self.cross = CrossBlock()
225
- self.glu = SwiGLU(d_model, 512)
226
 
227
  def _ema_over_time(self, score, alpha_dynamic):
228
  seq = tf.transpose(score, perm=[1, 0, 2])
 
196
  def __init__(self, d_model, num_heads, dff, dropout=0.1):
197
  super().__init__()
198
  self.mha = layers.MultiHeadAttention(num_heads=num_heads, key_dim=d_model)
199
+ self.ffn = SwiGLU(d_model, 320)
200
  self.norm1 = layers.LayerNormalization(epsilon=1e-6)
201
  self.norm2 = layers.LayerNormalization(epsilon=1e-6)
202
  self.dropout1 = layers.Dropout(dropout)
 
222
  self.alpha_linear = layers.Dense(1, activation='sigmoid', dtype='float32')
223
 
224
  self.cross = CrossBlock()
225
+ self.glu = SwiGLU(d_model, 320)
226
 
227
  def _ema_over_time(self, score, alpha_dynamic):
228
  seq = tf.transpose(score, perm=[1, 0, 2])