WCNegentropy's picture
๐Ÿš€ Refined BitTransformerLM: Organized codebase with best practices
b08919a verified
raw
history blame
3.47 kB
"""BitTransformerLM: Bit-native transformer with reversible layers and telemetry."""
# Core model components
from .model import (
BitTransformerLM,
PositionalEncoding,
ReversibleLoggingTransformerEncoderLayer,
diffusion_inference,
example_training_step,
example_usage,
infer_long_sequence,
)
# I/O and data processing
from .bit_io import bits_to_text, infer_text, text_to_bits
from .compression import (
compress_bits,
decompress_bits,
model_output_decompress,
pack_bits,
unpack_bits,
)
from .parity import enforce_parity
# Training and optimization
from .optimization import adjust_learning_rate, configure_optimizer
from .training import train_loop
# Model scaling and distillation
from .collapse import collapse_submodel, save_distilled_model
from .distil import TelemetryLog, distill_step
from .scale import expand_model
# Distributed computing
from .distributed import make_pipeline, wrap_fsdp
# Quantization support
from .quantization import convert_qat_fx, prepare_qat_fx, quantize_dynamic
# Safety and monitoring
from .safety import demo_hil_safety, hil_safe_inference, safe_sample_with_retry
from .telemetry import TelemetrySynthesizer, detect_metric_drift
# Configuration management
from .config import (
DataConfig,
ExperimentConfig,
ModelConfig,
SafetyConfig,
TrainingConfig,
get_config_from_env,
get_large_config,
get_medium_config,
get_small_config,
)
# Command-line interface
from .cli import dashboard_cli, infer_cli, train_cli
from .cli_standards import BitTransformerCLI
# Visualization and utilities
from .dashboard import plot_telemetry
from .dashboard_app import run_dashboard
from .hf_checkpoint import download_checkpoint, hf_login, save_checkpoint
from .torch_utils import cpu_autocast
from .utils import load_model, save_model, set_dropout
__all__ = [
# Core model components
"BitTransformerLM",
"PositionalEncoding",
"ReversibleLoggingTransformerEncoderLayer",
"diffusion_inference",
"example_training_step",
"example_usage",
"infer_long_sequence",
# I/O and data processing
"bits_to_text",
"compress_bits",
"decompress_bits",
"enforce_parity",
"infer_text",
"model_output_decompress",
"pack_bits",
"text_to_bits",
"unpack_bits",
# Training and optimization
"adjust_learning_rate",
"configure_optimizer",
"train_loop",
# Model scaling and distillation
"collapse_submodel",
"distill_step",
"expand_model",
"save_distilled_model",
"TelemetryLog",
# Distributed computing
"make_pipeline",
"wrap_fsdp",
# Quantization support
"convert_qat_fx",
"prepare_qat_fx",
"quantize_dynamic",
# Safety and monitoring
"demo_hil_safety",
"detect_metric_drift",
"hil_safe_inference",
"safe_sample_with_retry",
"TelemetrySynthesizer",
# Configuration management
"DataConfig",
"ExperimentConfig",
"get_config_from_env",
"get_large_config",
"get_medium_config",
"get_small_config",
"ModelConfig",
"SafetyConfig",
"TrainingConfig",
# Command-line interface
"BitTransformerCLI",
"dashboard_cli",
"infer_cli",
"train_cli",
# Visualization and utilities
"cpu_autocast",
"download_checkpoint",
"hf_login",
"load_model",
"plot_telemetry",
"run_dashboard",
"save_checkpoint",
"save_model",
"set_dropout",
]