Update modeling_custom_seq2seq_llm.py
Browse files
modeling_custom_seq2seq_llm.py
CHANGED
|
@@ -3,7 +3,7 @@ import torch.nn as nn
|
|
| 3 |
from torch.nn import CrossEntropyLoss
|
| 4 |
from transformers.modeling_outputs import Seq2SeqLMOutput
|
| 5 |
from transformers.activations import ACT2FN
|
| 6 |
-
from
|
| 7 |
from liger_kernel.transformers.cross_entropy import LigerCrossEntropyLoss
|
| 8 |
from liger_kernel.transformers.rms_norm import LigerRMSNorm
|
| 9 |
from liger_kernel.transformers.swiglu import LigerSwiGLUMLP
|
|
|
|
| 3 |
from torch.nn import CrossEntropyLoss
|
| 4 |
from transformers.modeling_outputs import Seq2SeqLMOutput
|
| 5 |
from transformers.activations import ACT2FN
|
| 6 |
+
from flash_atten import MHA # Import the MHA class from the provided implementation
|
| 7 |
from liger_kernel.transformers.cross_entropy import LigerCrossEntropyLoss
|
| 8 |
from liger_kernel.transformers.rms_norm import LigerRMSNorm
|
| 9 |
from liger_kernel.transformers.swiglu import LigerSwiGLUMLP
|