Commit ·
576edf6
1
Parent(s): 32a0629
Upload roberta_layers.py
Browse files- roberta_layers.py +3 -0
roberta_layers.py
CHANGED
|
@@ -199,6 +199,9 @@ class RobertaSelfAttention(nn.Module):
|
|
| 199 |
query_layer = self.transpose_for_scores(mixed_query_layer)
|
| 200 |
|
| 201 |
if xformers_available:
|
|
|
|
|
|
|
|
|
|
| 202 |
context_layer = xops.memory_efficient_attention(
|
| 203 |
query_layer, key_layer, value_layer, p=self.dropout_prob
|
| 204 |
)
|
|
|
|
| 199 |
query_layer = self.transpose_for_scores(mixed_query_layer)
|
| 200 |
|
| 201 |
if xformers_available:
|
| 202 |
+
query_layer = query_layer.transpose(1, 2)
|
| 203 |
+
key_layer = key_layer.transpose(1, 2)
|
| 204 |
+
value_layer = value_layer.transpose(1, 2)
|
| 205 |
context_layer = xops.memory_efficient_attention(
|
| 206 |
query_layer, key_layer, value_layer, p=self.dropout_prob
|
| 207 |
)
|