Yuchan
commited on
Update AlphaS2S.py
Browse files- AlphaS2S.py +2 -2
AlphaS2S.py
CHANGED
|
@@ -196,7 +196,7 @@ class EncoderBlock(layers.Layer):
|
|
| 196 |
def __init__(self, d_model, num_heads, dff, dropout=0.1):
|
| 197 |
super().__init__()
|
| 198 |
self.mha = layers.MultiHeadAttention(num_heads=num_heads, key_dim=d_model)
|
| 199 |
-
self.ffn = SwiGLU(d_model,
|
| 200 |
self.norm1 = layers.LayerNormalization(epsilon=1e-6)
|
| 201 |
self.norm2 = layers.LayerNormalization(epsilon=1e-6)
|
| 202 |
self.dropout1 = layers.Dropout(dropout)
|
|
@@ -222,7 +222,7 @@ class LoU(layers.Layer):
|
|
| 222 |
self.alpha_linear = layers.Dense(1, activation='sigmoid', dtype='float32')
|
| 223 |
|
| 224 |
self.cross = CrossBlock()
|
| 225 |
-
self.glu = SwiGLU(d_model,
|
| 226 |
|
| 227 |
def _ema_over_time(self, score, alpha_dynamic):
|
| 228 |
seq = tf.transpose(score, perm=[1, 0, 2])
|
|
|
|
| 196 |
def __init__(self, d_model, num_heads, dff, dropout=0.1):
|
| 197 |
super().__init__()
|
| 198 |
self.mha = layers.MultiHeadAttention(num_heads=num_heads, key_dim=d_model)
|
| 199 |
+
self.ffn = SwiGLU(d_model, 320)
|
| 200 |
self.norm1 = layers.LayerNormalization(epsilon=1e-6)
|
| 201 |
self.norm2 = layers.LayerNormalization(epsilon=1e-6)
|
| 202 |
self.dropout1 = layers.Dropout(dropout)
|
|
|
|
| 222 |
self.alpha_linear = layers.Dense(1, activation='sigmoid', dtype='float32')
|
| 223 |
|
| 224 |
self.cross = CrossBlock()
|
| 225 |
+
self.glu = SwiGLU(d_model, 320)
|
| 226 |
|
| 227 |
def _ema_over_time(self, score, alpha_dynamic):
|
| 228 |
seq = tf.transpose(score, perm=[1, 0, 2])
|