"""Model architecture module — LLaMA-style Decoder-Only Transformer.""" from .norm import RMSNorm from .rope import RotaryPositionalEmbedding from .attention import GroupedQueryAttention from .feedforward import SwiGLUFeedForward from .transformer_block import TransformerBlock from .llm_model import LLMModel from .utils import count_parameters_detailed, estimate_memory_gb __all__ = [ "RMSNorm", "RotaryPositionalEmbedding", "GroupedQueryAttention", "SwiGLUFeedForward", "TransformerBlock", "LLMModel", "count_parameters_detailed", "estimate_memory_gb", ]