| |
|
|
| from fla.layers import (ABCAttention, BasedLinearAttention, DeltaNet, |
| GatedLinearAttention, HGRN2Attention, LinearAttention, |
| MultiScaleRetention, ReBasedLinearAttention) |
| from fla.models import (ABCForCausalLM, ABCModel, DeltaNetForCausalLM, |
| DeltaNetModel, GLAForCausalLM, GLAModel, |
| HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM, |
| HGRNModel, LinearAttentionForCausalLM, |
| LinearAttentionModel, RetNetForCausalLM, RetNetModel, |
| RWKV6ForCausalLM, RWKV6Model, TransformerForCausalLM, |
| TransformerModel) |
| from fla.ops import (chunk_gla, chunk_retention, fused_chunk_based, |
| fused_chunk_gla, fused_chunk_retention) |
|
|
| __all__ = [ |
| 'ABCAttention', |
| 'BasedLinearAttention', |
| 'DeltaNet', |
| 'HGRN2Attention', |
| 'GatedLinearAttention', |
| 'LinearAttention', |
| 'MultiScaleRetention', |
| 'ReBasedLinearAttention', |
| 'ABCForCausalLM', |
| 'ABCModel', |
| 'DeltaNetForCausalLM', |
| 'DeltaNetModel', |
| 'HGRNForCausalLM', |
| 'HGRNModel', |
| 'HGRN2ForCausalLM', |
| 'HGRN2Model', |
| 'GLAForCausalLM', |
| 'GLAModel', |
| 'LinearAttentionForCausalLM', |
| 'LinearAttentionModel', |
| 'RetNetForCausalLM', |
| 'RetNetModel', |
| 'RWKV6ForCausalLM', |
| 'RWKV6Model', |
| 'TransformerForCausalLM', |
| 'TransformerModel', |
| 'chunk_gla', |
| 'chunk_retention', |
| 'fused_chunk_based', |
| 'fused_chunk_gla', |
| 'fused_chunk_retention' |
| ] |
|
|
| __version__ = '0.1' |
|
|