| """Model architecture module — LLaMA-style Decoder-Only Transformer.""" | |
| from .norm import RMSNorm | |
| from .rope import RotaryPositionalEmbedding | |
| from .attention import GroupedQueryAttention | |
| from .feedforward import SwiGLUFeedForward | |
| from .transformer_block import TransformerBlock | |
| from .llm_model import LLMModel | |
| from .utils import count_parameters_detailed, estimate_memory_gb | |
| __all__ = [ | |
| "RMSNorm", "RotaryPositionalEmbedding", "GroupedQueryAttention", | |
| "SwiGLUFeedForward", "TransformerBlock", "LLMModel", | |
| "count_parameters_detailed", "estimate_memory_gb", | |
| ] | |