Update modeling_utu
Browse files- modeling_utu.py +0 -1
modeling_utu.py
CHANGED
|
@@ -57,7 +57,6 @@ from .configuration_utu import UTUConfig
|
|
| 57 |
|
| 58 |
if is_torch_flex_attn_available():
|
| 59 |
from torch.nn.attention.flex_attention import BlockMask
|
| 60 |
-
|
| 61 |
from transformers.integrations.flex_attention import make_flex_block_causal_mask
|
| 62 |
|
| 63 |
|
|
|
|
| 57 |
|
| 58 |
if is_torch_flex_attn_available():
|
| 59 |
from torch.nn.attention.flex_attention import BlockMask
|
|
|
|
| 60 |
from transformers.integrations.flex_attention import make_flex_block_causal_mask
|
| 61 |
|
| 62 |
|