LLM-1B-Lab / llm_lab /model /__init__.py
Vjeong's picture
docs: translate all Korean comments and docstrings to English
858e8b2
raw
history blame contribute delete
573 Bytes
"""Model architecture module — LLaMA-style Decoder-Only Transformer."""
from .norm import RMSNorm
from .rope import RotaryPositionalEmbedding
from .attention import GroupedQueryAttention
from .feedforward import SwiGLUFeedForward
from .transformer_block import TransformerBlock
from .llm_model import LLMModel
from .utils import count_parameters_detailed, estimate_memory_gb
__all__ = [
"RMSNorm", "RotaryPositionalEmbedding", "GroupedQueryAttention",
"SwiGLUFeedForward", "TransformerBlock", "LLMModel",
"count_parameters_detailed", "estimate_memory_gb",
]