| | import json |
| | import os |
| | from typing import Any |
| |
|
| | import hydra |
| | import torch |
| | from hydra.utils import instantiate |
| | from accelerate.logging import get_logger |
| | from omegaconf import DictConfig, OmegaConf |
| | from torch import nn |
| |
|
| | from trainer.accelerators.base_accelerator import BaseAccelerator |
| | from trainer.configs.configs import TrainerConfig, instantiate_with_cfg |
| |
|
| | logger = get_logger(__name__) |
| |
|
| | os.environ["TOKENIZERS_PARALLELISM"] = "false" |
| |
|
| | def load_dataloaders(cfg: DictConfig) -> Any: |
| | dataloaders = {} |
| | for split in [cfg.train_split_name, cfg.valid_split_name, cfg.test_split_name]: |
| | dataset = instantiate_with_cfg(cfg, split=split) |
| | should_shuffle = split == cfg.train_split_name |
| | dataloaders[split] = torch.utils.data.DataLoader( |
| | dataset, |
| | shuffle=should_shuffle, |
| | batch_size=cfg.batch_size, |
| | collate_fn=dataset.collate_fn, |
| | num_workers=cfg.num_workers |
| | ) |
| | return dataloaders |
| |
|
| |
|
| | def load_optimizer(cfg: DictConfig, model: nn.Module): |
| | optimizer = instantiate(cfg, model=model) |
| | return optimizer |
| |
|
| |
|
| | def load_scheduler(cfg: DictConfig, optimizer): |
| | scheduler = instantiate_with_cfg(cfg, optimizer=optimizer) |
| | return scheduler |
| |
|
| |
|
| | def load_task(cfg: DictConfig, accelerator: BaseAccelerator): |
| | task = instantiate_with_cfg(cfg, accelerator=accelerator) |
| | return task |
| |
|
| |
|
| | def verify_or_write_config(cfg: TrainerConfig): |
| | os.makedirs(cfg.output_dir, exist_ok=True) |
| | yaml_path = os.path.join(cfg.output_dir, "config.yaml") |
| | if not os.path.exists(yaml_path): |
| | OmegaConf.save(cfg, yaml_path, resolve=True) |
| | with open(yaml_path) as f: |
| | existing_config = f.read() |
| | if existing_config != OmegaConf.to_yaml(cfg, resolve=True): |
| | raise ValueError(f"Config was not saved correctly - {yaml_path}") |
| | logger.info(f"Config can be found in {yaml_path}") |
| |
|
| |
|
| | @hydra.main(version_base=None, config_path="../conf", config_name="config") |
| | def main(cfg: TrainerConfig) -> None: |
| | accelerator = instantiate_with_cfg(cfg.accelerator) |
| |
|
| | if cfg.debug.activate and accelerator.is_main_process: |
| | import pydevd_pycharm |
| | pydevd_pycharm.settrace('localhost', port=cfg.debug.port, stdoutToServer=True, stderrToServer=True) |
| |
|
| | if accelerator.is_main_process: |
| | verify_or_write_config(cfg) |
| |
|
| | logger.info(f"Loading task") |
| | task = load_task(cfg.task, accelerator) |
| | logger.info(f"Loading model") |
| | model = instantiate_with_cfg(cfg.model) |
| | logger.info(f"Loading criterion") |
| | criterion = instantiate_with_cfg(cfg.criterion) |
| | logger.info(f"Loading optimizer") |
| | optimizer = load_optimizer(cfg.optimizer, model) |
| | logger.info(f"Loading lr scheduler") |
| | lr_scheduler = load_scheduler(cfg.lr_scheduler, optimizer) |
| | logger.info(f"Loading dataloaders") |
| | split2dataloader = load_dataloaders(cfg.dataset) |
| |
|
| | dataloaders = list(split2dataloader.values()) |
| | model, optimizer, lr_scheduler, *dataloaders = accelerator.prepare(model, optimizer, lr_scheduler, *dataloaders) |
| | split2dataloader = dict(zip(split2dataloader.keys(), dataloaders)) |
| |
|
| | accelerator.load_state_if_needed() |
| |
|
| | accelerator.recalc_train_length_after_prepare(len(split2dataloader[cfg.dataset.train_split_name])) |
| |
|
| | accelerator.init_training(cfg) |
| |
|
| | def evaluate(): |
| | model.eval() |
| | end_of_train_dataloader = accelerator.gradient_state.end_of_dataloader |
| | logger.info(f"*** Evaluating {cfg.dataset.valid_split_name} ***") |
| | metrics = task.evaluate(model, criterion, split2dataloader[cfg.dataset.valid_split_name]) |
| | accelerator.update_metrics(metrics) |
| | accelerator.gradient_state.end_of_dataloader = end_of_train_dataloader |
| |
|
| | logger.info(f"task: {task.__class__.__name__}") |
| | logger.info(f"model: {model.__class__.__name__}") |
| | logger.info(f"num. model params: {int(sum(p.numel() for p in model.parameters()) // 1e6)}M") |
| | logger.info( |
| | f"num. model trainable params: {int(sum(p.numel() for p in model.parameters() if p.requires_grad) // 1e6)}M") |
| | logger.info(f"criterion: {criterion.__class__.__name__}") |
| | logger.info(f"num. train examples: {len(split2dataloader[cfg.dataset.train_split_name].dataset)}") |
| | logger.info(f"num. valid examples: {len(split2dataloader[cfg.dataset.valid_split_name].dataset)}") |
| | logger.info(f"num. test examples: {len(split2dataloader[cfg.dataset.test_split_name].dataset)}") |
| |
|
| | for epoch in range(accelerator.cfg.num_epochs): |
| | train_loss, lr = 0.0, 0.0 |
| | for step, batch in enumerate(split2dataloader[cfg.dataset.train_split_name]): |
| | if accelerator.should_skip(epoch, step): |
| | accelerator.update_progbar_step() |
| | continue |
| |
|
| | if accelerator.should_eval(): |
| | evaluate() |
| |
|
| | if accelerator.should_save(): |
| | accelerator.save_checkpoint() |
| |
|
| | model.train() |
| |
|
| | with accelerator.accumulate(model): |
| | loss = task.train_step(model, criterion, batch) |
| | avg_loss = accelerator.gather(loss).mean().item() |
| |
|
| | accelerator.backward(loss) |
| | if accelerator.sync_gradients: |
| | accelerator.clip_grad_norm_(model.parameters()) |
| |
|
| | optimizer.step() |
| | lr_scheduler.step() |
| | optimizer.zero_grad() |
| |
|
| | train_loss += avg_loss / accelerator.cfg.gradient_accumulation_steps |
| |
|
| | if accelerator.sync_gradients: |
| | accelerator.update_global_step(train_loss) |
| | train_loss = 0.0 |
| |
|
| | if accelerator.global_step > 0: |
| | lr = lr_scheduler.get_last_lr()[0] |
| |
|
| | accelerator.update_step(avg_loss, lr) |
| |
|
| | if accelerator.should_end(): |
| | evaluate() |
| | accelerator.save_checkpoint() |
| | break |
| |
|
| | if accelerator.should_end(): |
| | break |
| |
|
| | accelerator.update_epoch() |
| |
|
| | accelerator.wait_for_everyone() |
| | accelerator.load_best_checkpoint() |
| | logger.info(f"*** Evaluating {cfg.dataset.valid_split_name} ***") |
| | metrics = task.evaluate(model, criterion, split2dataloader[cfg.dataset.valid_split_name]) |
| | accelerator.update_metrics(metrics) |
| | logger.info(f"*** Evaluating {cfg.dataset.test_split_name} ***") |
| | metrics = task.evaluate(model, criterion, split2dataloader[cfg.dataset.test_split_name]) |
| | metrics = {f"{cfg.dataset.test_split_name}_{k}": v for k, v in metrics.items()} |
| | accelerator.update_metrics(metrics) |
| | accelerator.unwrap_and_save(model) |
| | accelerator.end_training() |
| |
|
| |
|
| | if __name__ == '__main__': |
| | main() |
| |
|