hmr-dataset / tools /train.py
zirobtc's picture
Upload folder using huggingface_hub
7466fd2 verified
import hydra
import pytorch_lightning as pl
from omegaconf import DictConfig, OmegaConf
from pytorch_lightning.callbacks.checkpoint import Checkpoint
from hmr4d.utils.pylogger import Log
from hmr4d.configs import register_store_gvhmr
from hmr4d.utils.vis.rich_logger import print_cfg
from hmr4d.utils.net_utils import load_pretrained_model, get_resume_ckpt_path
def get_callbacks(cfg: DictConfig) -> list:
"""Parse and instantiate all the callbacks in the config."""
if not hasattr(cfg, "callbacks") or cfg.callbacks is None:
return None
# Handle special callbacks
enable_checkpointing = cfg.pl_trainer.get("enable_checkpointing", True)
# Instantiate all the callbacks
callbacks = []
for callback in cfg.callbacks.values():
if callback is not None:
cb = hydra.utils.instantiate(callback, _recursive_=False)
# skip when disable checkpointing and the callback is Checkpoint
if not enable_checkpointing and isinstance(cb, Checkpoint):
continue
else:
callbacks.append(cb)
return callbacks
def train(cfg: DictConfig) -> None:
"""Train/Test"""
Log.info(f"[Exp Name]: {cfg.exp_name}")
if cfg.task == "fit":
Log.info(f"[GPU x Batch] = {cfg.pl_trainer.devices} x {cfg.data.loader_opts.train.batch_size}")
pl.seed_everything(cfg.seed)
# preparation
datamodule: pl.LightningDataModule = hydra.utils.instantiate(cfg.data, _recursive_=False)
model: pl.LightningModule = hydra.utils.instantiate(cfg.model, _recursive_=False)
if cfg.ckpt_path is not None:
load_pretrained_model(model, cfg.ckpt_path)
# PL callbacks and logger
callbacks = get_callbacks(cfg)
has_ckpt_cb = any([isinstance(cb, Checkpoint) for cb in callbacks])
if not has_ckpt_cb and cfg.pl_trainer.get("enable_checkpointing", True):
Log.warning("No checkpoint-callback found. Disabling PL auto checkpointing.")
cfg.pl_trainer = {**cfg.pl_trainer, "enable_checkpointing": False}
logger = hydra.utils.instantiate(cfg.logger, _recursive_=False)
# PL-Trainer
if cfg.task == "test":
Log.info("Test mode forces full-precision.")
cfg.pl_trainer = {**cfg.pl_trainer, "precision": 32}
trainer = pl.Trainer(
accelerator="gpu",
logger=logger if logger is not None else False,
callbacks=callbacks,
**cfg.pl_trainer,
)
if cfg.task == "fit":
resume_path = None
if cfg.resume_mode is not None:
resume_path = get_resume_ckpt_path(cfg.resume_mode, ckpt_dir=cfg.callbacks.model_checkpoint.dirpath)
Log.info(f"Resume training from {resume_path}")
Log.info("Start Fitiing...")
trainer.fit(model, datamodule.train_dataloader(), datamodule.val_dataloader(), ckpt_path=resume_path)
elif cfg.task == "test":
Log.info("Start Testing...")
trainer.test(model, datamodule.test_dataloader())
else:
raise ValueError(f"Unknown task: {cfg.task}")
Log.info("End of script.")
@hydra.main(version_base="1.3", config_path="../hmr4d/configs", config_name="train")
def main(cfg) -> None:
print_cfg(cfg, use_rich=True)
train(cfg)
if __name__ == "__main__":
register_store_gvhmr()
main()