Update modeling_custom_seq2seq_llm.py
Browse files
modeling_custom_seq2seq_llm.py
CHANGED
|
@@ -7,6 +7,7 @@ from flash_atten import MHA # Import the MHA class from the provided implementa
|
|
| 7 |
from liger_kernel.transformers.cross_entropy import LigerCrossEntropyLoss
|
| 8 |
from liger_kernel.transformers.rms_norm import LigerRMSNorm
|
| 9 |
from liger_kernel.transformers.swiglu import LigerSwiGLUMLP
|
|
|
|
| 10 |
|
| 11 |
from configuration_custom_seq2seq_llm import Seq2SeqConfig
|
| 12 |
|
|
|
|
| 7 |
from liger_kernel.transformers.cross_entropy import LigerCrossEntropyLoss
|
| 8 |
from liger_kernel.transformers.rms_norm import LigerRMSNorm
|
| 9 |
from liger_kernel.transformers.swiglu import LigerSwiGLUMLP
|
| 10 |
+
from transformers import PreTrainedModel
|
| 11 |
|
| 12 |
from configuration_custom_seq2seq_llm import Seq2SeqConfig
|
| 13 |
|