gated_deltaproduct / fla2 /__init__.py
msj19's picture
Add files using upload-large-folder tool
e811e56 verified
# -*- coding: utf-8 -*-
# from fla.layers import (ABCAttention, Attention, BasedLinearAttention,
# DeltaNet, GatedLinearAttention, HGRN2Attention,
# LinearAttention, MultiScaleRetention,
# ReBasedLinearAttention)
# from fla.models import (ABCForCausalLM, ABCModel, DeltaNetForCausalLM,
# DeltaNetModel, GLAForCausalLM, GLAModel,
# HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM,
# HGRNModel, LinearAttentionForCausalLM,
# LinearAttentionModel, RetNetForCausalLM, RetNetModel,
# RWKV6ForCausalLM, RWKV6Model, TransformerForCausalLM,
# TransformerModel)
# from fla.ops import (chunk_gla, chunk_retention, fused_chunk_based,
# fused_chunk_gla, fused_chunk_retention)
from .models import emla,emgla,mask_deltanet,mask_gdn,transformer
# __all__ = [
# 'ABCAttention',
# 'Attention',
# 'BasedLinearAttention',
# 'DeltaNet',
# 'HGRN2Attention',
# 'GatedLinearAttention',
# 'LinearAttention',
# 'MultiScaleRetention',
# 'ReBasedLinearAttention',
# 'ABCForCausalLM',
# 'ABCModel',
# 'DeltaNetForCausalLM',
# 'DeltaNetModel',
# 'HGRNForCausalLM',
# 'HGRNModel',
# 'HGRN2ForCausalLM',
# 'HGRN2Model',
# 'GLAForCausalLM',
# 'GLAModel',
# 'LinearAttentionForCausalLM',
# 'LinearAttentionModel',
# 'RetNetForCausalLM',
# 'RetNetModel',
# 'RWKV6ForCausalLM',
# 'RWKV6Model',
# 'TransformerForCausalLM',
# 'TransformerModel',
# 'chunk_gla',
# 'chunk_retention',
# 'fused_chunk_based',
# 'fused_chunk_gla',
# 'fused_chunk_retention'
# ]
__version__ = '0.1'