| | |
| | """ Checkpoint Averaging Script |
| | |
| | This script averages all model weights for checkpoints in specified path that match |
| | the specified filter wildcard. All checkpoints must be from the exact same model. |
| | |
| | For any hope of decent results, the checkpoints should be from the same or child |
| | (via resumes) training session. This can be viewed as similar to maintaining running |
| | EMA (exponential moving average) of the model weights or performing SWA (stochastic |
| | weight averaging), but post-training. |
| | |
| | Hacked together by / Copyright 2020 Ross Wightman (https://github.com/rwightman) |
| | """ |
| | import torch |
| | import argparse |
| | import os |
| | import glob |
| | import hashlib |
| | from timm.models import load_state_dict |
| | try: |
| | import safetensors.torch |
| | _has_safetensors = True |
| | except ImportError: |
| | _has_safetensors = False |
| |
|
| | DEFAULT_OUTPUT = "./averaged.pth" |
| | DEFAULT_SAFE_OUTPUT = "./averaged.safetensors" |
| |
|
| | parser = argparse.ArgumentParser(description='PyTorch Checkpoint Averager') |
| | parser.add_argument('--input', default='', type=str, metavar='PATH', |
| | help='path to base input folder containing checkpoints') |
| | parser.add_argument('--filter', default='*.pth.tar', type=str, metavar='WILDCARD', |
| | help='checkpoint filter (path wildcard)') |
| | parser.add_argument('--output', default=DEFAULT_OUTPUT, type=str, metavar='PATH', |
| | help=f'Output filename. Defaults to {DEFAULT_SAFE_OUTPUT} when passing --safetensors.') |
| | parser.add_argument('--no-use-ema', dest='no_use_ema', action='store_true', |
| | help='Force not using ema version of weights (if present)') |
| | parser.add_argument('--no-sort', dest='no_sort', action='store_true', |
| | help='Do not sort and select by checkpoint metric, also makes "n" argument irrelevant') |
| | parser.add_argument('-n', type=int, default=10, metavar='N', |
| | help='Number of checkpoints to average') |
| | parser.add_argument('--safetensors', action='store_true', |
| | help='Save weights using safetensors instead of the default torch way (pickle).') |
| |
|
| |
|
| | def checkpoint_metric(checkpoint_path): |
| | if not checkpoint_path or not os.path.isfile(checkpoint_path): |
| | return {} |
| | print("=> Extracting metric from checkpoint '{}'".format(checkpoint_path)) |
| | checkpoint = torch.load(checkpoint_path, map_location='cpu') |
| | metric = None |
| | if 'metric' in checkpoint: |
| | metric = checkpoint['metric'] |
| | elif 'metrics' in checkpoint and 'metric_name' in checkpoint: |
| | metrics = checkpoint['metrics'] |
| | print(metrics) |
| | metric = metrics[checkpoint['metric_name']] |
| | return metric |
| |
|
| |
|
| | def main(): |
| | args = parser.parse_args() |
| | |
| | args.use_ema = not args.no_use_ema |
| | |
| | args.sort = not args.no_sort |
| |
|
| | if args.safetensors and args.output == DEFAULT_OUTPUT: |
| | |
| | args.output = DEFAULT_SAFE_OUTPUT |
| |
|
| | output, output_ext = os.path.splitext(args.output) |
| | if not output_ext: |
| | output_ext = ('.safetensors' if args.safetensors else '.pth') |
| | output = output + output_ext |
| |
|
| | if args.safetensors and not output_ext == ".safetensors": |
| | print( |
| | "Warning: saving weights as safetensors but output file extension is not " |
| | f"set to '.safetensors': {args.output}" |
| | ) |
| |
|
| | if os.path.exists(output): |
| | print("Error: Output filename ({}) already exists.".format(output)) |
| | exit(1) |
| |
|
| | pattern = args.input |
| | if not args.input.endswith(os.path.sep) and not args.filter.startswith(os.path.sep): |
| | pattern += os.path.sep |
| | pattern += args.filter |
| | checkpoints = glob.glob(pattern, recursive=True) |
| |
|
| | if args.sort: |
| | checkpoint_metrics = [] |
| | for c in checkpoints: |
| | metric = checkpoint_metric(c) |
| | if metric is not None: |
| | checkpoint_metrics.append((metric, c)) |
| | checkpoint_metrics = list(sorted(checkpoint_metrics)) |
| | checkpoint_metrics = checkpoint_metrics[-args.n:] |
| | if checkpoint_metrics: |
| | print("Selected checkpoints:") |
| | [print(m, c) for m, c in checkpoint_metrics] |
| | avg_checkpoints = [c for m, c in checkpoint_metrics] |
| | else: |
| | avg_checkpoints = checkpoints |
| | if avg_checkpoints: |
| | print("Selected checkpoints:") |
| | [print(c) for c in checkpoints] |
| |
|
| | if not avg_checkpoints: |
| | print('Error: No checkpoints found to average.') |
| | exit(1) |
| |
|
| | avg_state_dict = {} |
| | avg_counts = {} |
| | for c in avg_checkpoints: |
| | new_state_dict = load_state_dict(c, args.use_ema) |
| | if not new_state_dict: |
| | print(f"Error: Checkpoint ({c}) doesn't exist") |
| | continue |
| | for k, v in new_state_dict.items(): |
| | if k not in avg_state_dict: |
| | avg_state_dict[k] = v.clone().to(dtype=torch.float64) |
| | avg_counts[k] = 1 |
| | else: |
| | avg_state_dict[k] += v.to(dtype=torch.float64) |
| | avg_counts[k] += 1 |
| |
|
| | for k, v in avg_state_dict.items(): |
| | v.div_(avg_counts[k]) |
| |
|
| | |
| | float32_info = torch.finfo(torch.float32) |
| | final_state_dict = {} |
| | for k, v in avg_state_dict.items(): |
| | v = v.clamp(float32_info.min, float32_info.max) |
| | final_state_dict[k] = v.to(dtype=torch.float32) |
| |
|
| | if args.safetensors: |
| | assert _has_safetensors, "`pip install safetensors` to use .safetensors" |
| | safetensors.torch.save_file(final_state_dict, output) |
| | else: |
| | torch.save(final_state_dict, output) |
| |
|
| | with open(output, 'rb') as f: |
| | sha_hash = hashlib.sha256(f.read()).hexdigest() |
| | print(f"=> Saved state_dict to '{output}, SHA256: {sha_hash}'") |
| |
|
| |
|
| | if __name__ == '__main__': |
| | main() |
| |
|