proposed patch for the Tensor size mismatch error
#15
by
claverAI
- opened
- modeling_phi.py +1 -0
modeling_phi.py
CHANGED
|
@@ -315,6 +315,7 @@ class CrossAttention(nn.Module):
|
|
| 315 |
dtype=scores.dtype,
|
| 316 |
device=scores.device,
|
| 317 |
)
|
|
|
|
| 318 |
padding_mask.masked_fill_(key_padding_mask, 0.0)
|
| 319 |
scores = scores + rearrange(padding_mask, "b s -> b 1 1 s")
|
| 320 |
|
|
|
|
| 315 |
dtype=scores.dtype,
|
| 316 |
device=scores.device,
|
| 317 |
)
|
| 318 |
+
key_padding_mask = key_padding_mask[:, :seqlen_k]
|
| 319 |
padding_mask.masked_fill_(key_padding_mask, 0.0)
|
| 320 |
scores = scores + rearrange(padding_mask, "b s -> b 1 1 s")
|
| 321 |
|