adalbertojunior commited on
Commit
576edf6
·
1 Parent(s): 32a0629

Upload roberta_layers.py

Browse files
Files changed (1) hide show
  1. roberta_layers.py +3 -0
roberta_layers.py CHANGED
@@ -199,6 +199,9 @@ class RobertaSelfAttention(nn.Module):
199
  query_layer = self.transpose_for_scores(mixed_query_layer)
200
 
201
  if xformers_available:
 
 
 
202
  context_layer = xops.memory_efficient_attention(
203
  query_layer, key_layer, value_layer, p=self.dropout_prob
204
  )
 
199
  query_layer = self.transpose_for_scores(mixed_query_layer)
200
 
201
  if xformers_available:
202
+ query_layer = query_layer.transpose(1, 2)
203
+ key_layer = key_layer.transpose(1, 2)
204
+ value_layer = value_layer.transpose(1, 2)
205
  context_layer = xops.memory_efficient_attention(
206
  query_layer, key_layer, value_layer, p=self.dropout_prob
207
  )