msj19's picture
Add files using upload-large-folder tool
e73a905 verified
# # -*- coding: utf-8 -*-
# from fla.models.abc import ABCConfig, ABCForCausalLM, ABCModel
# from fla.models.bitnet import BitNetConfig, BitNetForCausalLM, BitNetModel
from .delta_net import DeltaNetConfig, DeltaNetForCausalLM, DeltaNetModel
# from fla.models.forgetting_transformer import (
# ForgettingTransformerConfig,
# ForgettingTransformerForCausalLM,
# ForgettingTransformerModel
# )
from .gated_deltanet import GatedDeltaNetConfig, GatedDeltaNetForCausalLM, GatedDeltaNetModel
# from fla.models.gated_deltaproduct import GatedDeltaProductConfig, GatedDeltaProductForCausalLM, GatedDeltaProductModel
# from fla.models.gla import GLAConfig, GLAForCausalLM, GLAModel
# from fla.models.gsa import GSAConfig, GSAForCausalLM, GSAModel
# from fla.models.hgrn import HGRNConfig, HGRNForCausalLM, HGRNModel
# from fla.models.hgrn2 import HGRN2Config, HGRN2ForCausalLM, HGRN2Model
# from fla.models.lightnet import LightNetConfig, LightNetForCausalLM, LightNetModel
# from fla.models.linear_attn import LinearAttentionConfig, LinearAttentionForCausalLM, LinearAttentionModel
# from fla.models.mamba import MambaConfig, MambaForCausalLM, MambaModel
# from fla.models.mamba2 import Mamba2Config, Mamba2ForCausalLM, Mamba2Model
# from fla.models.nsa import NSAConfig, NSAForCausalLM, NSAModel
# from fla.models.path_attn import PaTHAttentionConfig, PaTHAttentionForCausalLM, PaTHAttentionModel
# from fla.models.retnet import RetNetConfig, RetNetForCausalLM, RetNetModel
# from fla.models.rwkv6 import RWKV6Config, RWKV6ForCausalLM, RWKV6Model
# from fla.models.rwkv7 import RWKV7Config, RWKV7ForCausalLM, RWKV7Model
# from fla.models.samba import SambaConfig, SambaForCausalLM, SambaModel
# from fla.models.transformer import TransformerConfig, TransformerForCausalLM, TransformerModel
# __all__ = [
# 'ABCConfig', 'ABCForCausalLM', 'ABCModel',
# 'BitNetConfig', 'BitNetForCausalLM', 'BitNetModel',
# 'DeltaNetConfig', 'DeltaNetForCausalLM', 'DeltaNetModel',
# 'ForgettingTransformerConfig', 'ForgettingTransformerForCausalLM', 'ForgettingTransformerModel',
# 'GatedDeltaNetConfig', 'GatedDeltaNetForCausalLM', 'GatedDeltaNetModel',
# 'GLAConfig', 'GLAForCausalLM', 'GLAModel',
# 'GSAConfig', 'GSAForCausalLM', 'GSAModel',
# 'HGRNConfig', 'HGRNForCausalLM', 'HGRNModel',
# 'HGRN2Config', 'HGRN2ForCausalLM', 'HGRN2Model',
# 'LightNetConfig', 'LightNetForCausalLM', 'LightNetModel',
# 'LinearAttentionConfig', 'LinearAttentionForCausalLM', 'LinearAttentionModel',
# 'MambaConfig', 'MambaForCausalLM', 'MambaModel',
# 'Mamba2Config', 'Mamba2ForCausalLM', 'Mamba2Model',
# 'NSAConfig', 'NSAForCausalLM', 'NSAModel',
# 'RetNetConfig', 'RetNetForCausalLM', 'RetNetModel',
# 'RWKV6Config', 'RWKV6ForCausalLM', 'RWKV6Model',
# 'RWKV7Config', 'RWKV7ForCausalLM', 'RWKV7Model',
# 'SambaConfig', 'SambaForCausalLM', 'SambaModel',
# 'TransformerConfig', 'TransformerForCausalLM', 'TransformerModel',
# 'GatedDeltaProductConfig', 'GatedDeltaProductForCausalLM', 'GatedDeltaProductModel',
# 'PaTHAttentionConfig', 'PaTHAttentionForCausalLM', 'PaTHAttentionModel',
# ]
# from .emla import emlaConfig,emlaForCausalLM,emlaModel
# from .emgla import emglaConfig,emglaForCausalLM,emglaModel
from .emdeltanet import emdeltanetConfig,emdeltanetForCausalLM,emdeltanetModel
# from .transformer import TransformerConfig,TransformerForCausalLM,TransformerModel