| from .model import ( |
| PositionalEncoding, |
| BitTransformerLM, |
| ReversibleLoggingTransformerEncoderLayer, |
| example_usage, |
| example_training_step, |
| infer_long_sequence, |
| diffusion_inference, |
| ) |
| from .telemetry import TelemetrySynthesizer, detect_metric_drift |
| from .dashboard import plot_telemetry |
| from .dashboard_app import run_dashboard |
| from .collapse import collapse_submodel, save_distilled_model |
| from .safety import hil_safe_inference, demo_hil_safety, safe_sample_with_retry |
| from .bit_io import ( |
| text_to_bits, |
| bits_to_text, |
| infer_text, |
| ) |
| from .parity import enforce_parity |
| from .compression import ( |
| compress_bits, |
| decompress_bits, |
| model_output_decompress, |
| pack_bits, |
| unpack_bits, |
| ) |
| from .distributed import wrap_fsdp, make_pipeline |
| from .optimization import configure_optimizer, adjust_learning_rate |
| from .scale import expand_model |
| from .distil import distill_step, TelemetryLog |
| from .quantization import ( |
| quantize_dynamic, |
| prepare_qat_fx, |
| convert_qat_fx, |
| ) |
| from .training import train_loop |
| from .utils import save_model, load_model, set_dropout |
| from .hf_checkpoint import hf_login, save_checkpoint, download_checkpoint |
| from .torch_utils import cpu_autocast |
|
|
| __all__ = [ |
| "PositionalEncoding", |
| "BitTransformerLM", |
| "ReversibleLoggingTransformerEncoderLayer", |
| "example_usage", |
| "example_training_step", |
| "TelemetrySynthesizer", |
| "detect_metric_drift", |
| "collapse_submodel", |
| "save_distilled_model", |
| "hil_safe_inference", |
| "demo_hil_safety", |
| "safe_sample_with_retry", |
| "text_to_bits", |
| "bits_to_text", |
| "infer_text", |
| "enforce_parity", |
| "plot_telemetry", |
| "run_dashboard", |
| "configure_optimizer", |
| "adjust_learning_rate", |
| "expand_model", |
| "distill_step", |
| "TelemetryLog", |
| "quantize_dynamic", |
| "prepare_qat_fx", |
| "convert_qat_fx", |
| "train_loop", |
| "wrap_fsdp", |
| "make_pipeline", |
| "compress_bits", |
| "decompress_bits", |
| "model_output_decompress", |
| "pack_bits", |
| "unpack_bits", |
| "infer_long_sequence", |
| "diffusion_inference", |
| "save_model", |
| "load_model", |
| "set_dropout", |
| "hf_login", |
| "save_checkpoint", |
| "download_checkpoint", |
| "cpu_autocast", |
| ] |
|
|