;)
Browse files- modeling_t2.py +2 -0
modeling_t2.py
CHANGED
|
@@ -149,6 +149,8 @@ class TransformerAttention(nn.Module):
|
|
| 149 |
super().__init__()
|
| 150 |
self.config = config
|
| 151 |
self.stage = stage
|
|
|
|
|
|
|
| 152 |
self.head_dim = config.hidden_size // config.num_attention_heads
|
| 153 |
assert (
|
| 154 |
self.head_dim * config.num_attention_heads == config.hidden_size
|
|
|
|
| 149 |
super().__init__()
|
| 150 |
self.config = config
|
| 151 |
self.stage = stage
|
| 152 |
+
self.max_position_embeddings = config.max_position_embeddings
|
| 153 |
+
self.rope_theta = config.rope_theta
|
| 154 |
self.head_dim = config.hidden_size // config.num_attention_heads
|
| 155 |
assert (
|
| 156 |
self.head_dim * config.num_attention_heads == config.hidden_size
|