| """ Attention and normalization modules """ | |
| from onmt.modules.util_class import Elementwise | |
| from onmt.modules.gate import context_gate_factory, ContextGate | |
| from onmt.modules.global_attention import GlobalAttention | |
| from onmt.modules.conv_multi_step_attention import ConvMultiStepAttention | |
| from onmt.modules.copy_generator import CopyGenerator, CopyGeneratorLoss | |
| from onmt.modules.multi_headed_attn import MultiHeadedAttention | |
| from onmt.modules.embeddings import Embeddings, PositionalEncoding | |
| from onmt.modules.weight_norm import WeightNormConv2d | |
| from onmt.modules.average_attn import AverageAttention | |
| from onmt.modules.alibi_position_bias import AlibiPositionalBias | |
| from onmt.modules.rmsnorm import RMSNorm | |
| __all__ = [ | |
| "Elementwise", | |
| "context_gate_factory", | |
| "ContextGate", | |
| "GlobalAttention", | |
| "ConvMultiStepAttention", | |
| "CopyGenerator", | |
| "CopyGeneratorLoss", | |
| "CopyGeneratorLMLossCompute", | |
| "MultiHeadedAttention", | |
| "Embeddings", | |
| "PositionalEncoding", | |
| "AlibiPositionalBias", | |
| "WeightNormConv2d", | |
| "AverageAttention", | |
| "RMSNorm", | |
| ] | |