| # # -*- coding: utf-8 -*- | |
| # from fla.models.abc import ABCConfig, ABCForCausalLM, ABCModel | |
| # from fla.models.bitnet import BitNetConfig, BitNetForCausalLM, BitNetModel | |
| from .delta_net import DeltaNetConfig, DeltaNetForCausalLM, DeltaNetModel | |
| # from fla.models.forgetting_transformer import ( | |
| # ForgettingTransformerConfig, | |
| # ForgettingTransformerForCausalLM, | |
| # ForgettingTransformerModel | |
| # ) | |
| from .gated_deltanet import GatedDeltaNetConfig, GatedDeltaNetForCausalLM, GatedDeltaNetModel | |
| # from fla.models.gated_deltaproduct import GatedDeltaProductConfig, GatedDeltaProductForCausalLM, GatedDeltaProductModel | |
| # from fla.models.gla import GLAConfig, GLAForCausalLM, GLAModel | |
| # from fla.models.gsa import GSAConfig, GSAForCausalLM, GSAModel | |
| # from fla.models.hgrn import HGRNConfig, HGRNForCausalLM, HGRNModel | |
| # from fla.models.hgrn2 import HGRN2Config, HGRN2ForCausalLM, HGRN2Model | |
| # from fla.models.lightnet import LightNetConfig, LightNetForCausalLM, LightNetModel | |
| # from fla.models.linear_attn import LinearAttentionConfig, LinearAttentionForCausalLM, LinearAttentionModel | |
| # from fla.models.mamba import MambaConfig, MambaForCausalLM, MambaModel | |
| # from fla.models.mamba2 import Mamba2Config, Mamba2ForCausalLM, Mamba2Model | |
| # from fla.models.nsa import NSAConfig, NSAForCausalLM, NSAModel | |
| # from fla.models.path_attn import PaTHAttentionConfig, PaTHAttentionForCausalLM, PaTHAttentionModel | |
| # from fla.models.retnet import RetNetConfig, RetNetForCausalLM, RetNetModel | |
| # from fla.models.rwkv6 import RWKV6Config, RWKV6ForCausalLM, RWKV6Model | |
| # from fla.models.rwkv7 import RWKV7Config, RWKV7ForCausalLM, RWKV7Model | |
| # from fla.models.samba import SambaConfig, SambaForCausalLM, SambaModel | |
| # from fla.models.transformer import TransformerConfig, TransformerForCausalLM, TransformerModel | |
| # __all__ = [ | |
| # 'ABCConfig', 'ABCForCausalLM', 'ABCModel', | |
| # 'BitNetConfig', 'BitNetForCausalLM', 'BitNetModel', | |
| # 'DeltaNetConfig', 'DeltaNetForCausalLM', 'DeltaNetModel', | |
| # 'ForgettingTransformerConfig', 'ForgettingTransformerForCausalLM', 'ForgettingTransformerModel', | |
| # 'GatedDeltaNetConfig', 'GatedDeltaNetForCausalLM', 'GatedDeltaNetModel', | |
| # 'GLAConfig', 'GLAForCausalLM', 'GLAModel', | |
| # 'GSAConfig', 'GSAForCausalLM', 'GSAModel', | |
| # 'HGRNConfig', 'HGRNForCausalLM', 'HGRNModel', | |
| # 'HGRN2Config', 'HGRN2ForCausalLM', 'HGRN2Model', | |
| # 'LightNetConfig', 'LightNetForCausalLM', 'LightNetModel', | |
| # 'LinearAttentionConfig', 'LinearAttentionForCausalLM', 'LinearAttentionModel', | |
| # 'MambaConfig', 'MambaForCausalLM', 'MambaModel', | |
| # 'Mamba2Config', 'Mamba2ForCausalLM', 'Mamba2Model', | |
| # 'NSAConfig', 'NSAForCausalLM', 'NSAModel', | |
| # 'RetNetConfig', 'RetNetForCausalLM', 'RetNetModel', | |
| # 'RWKV6Config', 'RWKV6ForCausalLM', 'RWKV6Model', | |
| # 'RWKV7Config', 'RWKV7ForCausalLM', 'RWKV7Model', | |
| # 'SambaConfig', 'SambaForCausalLM', 'SambaModel', | |
| # 'TransformerConfig', 'TransformerForCausalLM', 'TransformerModel', | |
| # 'GatedDeltaProductConfig', 'GatedDeltaProductForCausalLM', 'GatedDeltaProductModel', | |
| # 'PaTHAttentionConfig', 'PaTHAttentionForCausalLM', 'PaTHAttentionModel', | |
| # ] | |
| # from .emla import emlaConfig,emlaForCausalLM,emlaModel | |
| # from .emgla import emglaConfig,emglaForCausalLM,emglaModel | |
| from .emdeltanet import emdeltanetConfig,emdeltanetForCausalLM,emdeltanetModel | |
| # from .transformer import TransformerConfig,TransformerForCausalLM,TransformerModel | |