EDM / src /config /default.py
lixi042
update
7e31006
from yacs.config import CfgNode as CN
_CN = CN()
############## ↓ EDM Pipeline ↓ ##############
_CN.EDM = CN()
_CN.EDM.TRAIN_RES_H = 832
_CN.EDM.TRAIN_RES_W = 832
_CN.EDM.TEST_RES_H = 832
_CN.EDM.TEST_RES_W = 832
_CN.EDM.LOCAL_RESOLUTION = 8 # coarse matching at 1/8, window_size 8x8 in fine_level
_CN.EDM.MP = False # Not use mixed precision (ELoFTR defualt used mixed precision)
_CN.EDM.HALF = False # Not FP16
_CN.EDM.DEPLOY = False # export onnx model
_CN.EDM.EVAL_TIMES = 1
# 3. Coarse-Matching config
_CN.EDM.BACKBONE = CN()
_CN.EDM.BACKBONE.BLOCK_DIMS = [32, 64, 128, 256, 256] # 1/2 -> 1/32
# 3. Coarse-Matching config
_CN.EDM.NECK = CN()
_CN.EDM.NECK.D_MODEL = 256
_CN.EDM.NECK.NHEAD = 8
_CN.EDM.NECK.LAYER_NAMES = ["self", "cross"] * 2
_CN.EDM.NECK.AGG_SIZE0 = 1
_CN.EDM.NECK.AGG_SIZE1 = 1
_CN.EDM.NECK.ROPE = True
_CN.EDM.NECK.NPE = None
# 3. Coarse-Matching config
_CN.EDM.COARSE = CN()
_CN.EDM.COARSE.MCONF_THR = 0.2
_CN.EDM.COARSE.BORDER_RM = 0
_CN.EDM.COARSE.DSMAX_TEMPERATURE = 0.1
_CN.EDM.COARSE.TRAIN_PAD_NUM = 32 # training tricks: avoid DDP deadlock
_CN.EDM.COARSE.TOPK = 2048
_CN.EDM.COARSE.DS_OPT = True
# 4. EDM-fine module config
_CN.EDM.FINE = CN()
_CN.EDM.FINE.DROPRATE = None
_CN.EDM.FINE.COORD_LENGTH = 16
_CN.EDM.FINE.BI_DIRECTIONAL_REFINE = True
_CN.EDM.FINE.SIGMA_THR = 0.0
_CN.EDM.FINE.SIGMA_SELECTION = True
# 5. EDM Losses
# -- # coarse-level
_CN.EDM.LOSS = CN()
_CN.EDM.LOSS.COARSE_TYPE = "focal"
_CN.EDM.LOSS.COARSE_WEIGHT = 1.0
_CN.EDM.LOSS.SPARSE_SPVS = True
# -- - -- # focal loss (coarse)
_CN.EDM.LOSS.FOCAL_ALPHA = 0.25
_CN.EDM.LOSS.FOCAL_GAMMA = 2.0
_CN.EDM.LOSS.POS_WEIGHT = 1.0
_CN.EDM.LOSS.NEG_WEIGHT = 1.0
# -- # fine-level
_CN.EDM.LOSS.FINE_TYPE = "rle"
_CN.EDM.LOSS.FINE_WEIGHT = 0.2
_CN.EDM.LOSS.Q_DISTRIBUTION = "laplace" # options: ['laplace', 'gaussian']
############## Dataset ##############
_CN.DATASET = CN()
# 1. data config
# training and validating
_CN.DATASET.TRAINVAL_DATA_SOURCE = None # options: ['ScanNet', 'MegaDepth']
_CN.DATASET.TRAIN_DATA_ROOT = None
_CN.DATASET.TRAIN_POSE_ROOT = None # (optional directory for poses)
_CN.DATASET.TRAIN_NPZ_ROOT = None
_CN.DATASET.TRAIN_LIST_PATH = None
_CN.DATASET.TRAIN_INTRINSIC_PATH = None
_CN.DATASET.VAL_DATA_ROOT = None
_CN.DATASET.VAL_POSE_ROOT = None # (optional directory for poses)
_CN.DATASET.VAL_NPZ_ROOT = None
_CN.DATASET.VAL_LIST_PATH = (
None # None if val data from all scenes are bundled into a single npz file
)
_CN.DATASET.VAL_INTRINSIC_PATH = None
# testing
_CN.DATASET.TEST_DATA_SOURCE = None
_CN.DATASET.TEST_DATA_ROOT = None
_CN.DATASET.TEST_POSE_ROOT = None # (optional directory for poses)
_CN.DATASET.TEST_NPZ_ROOT = None
_CN.DATASET.TEST_LIST_PATH = (
None # None if test data from all scenes are bundled into a single npz file
)
_CN.DATASET.TEST_INTRINSIC_PATH = None
# 2. dataset config
# general options
_CN.DATASET.MIN_OVERLAP_SCORE_TRAIN = (
0.4 # discard data with overlap_score < min_overlap_score
)
_CN.DATASET.MIN_OVERLAP_SCORE_TEST = 0.0
_CN.DATASET.AUGMENTATION_TYPE = None # options: [None, 'dark', 'mobile']
# MegaDepth options
_CN.DATASET.MGDPT_IMG_PAD = True # pad img to square with size = MGDPT_IMG_RESIZE
_CN.DATASET.MGDPT_DEPTH_PAD = True # pad depthmap to square with size = 2000
_CN.DATASET.MGDPT_DF = 8
############## Trainer ##############
_CN.TRAINER = CN()
_CN.TRAINER.WORLD_SIZE = 1
_CN.TRAINER.CANONICAL_BS = 4 * 8
_CN.TRAINER.CANONICAL_LR = 2e-3
_CN.TRAINER.SCALING = None # this will be calculated automatically
_CN.TRAINER.FIND_LR = False # use learning rate finder from pytorch-lightning
# optimizer
_CN.TRAINER.OPTIMIZER = "adamw" # [adam, adamw]
_CN.TRAINER.TRUE_LR = None # this will be calculated automatically at runtime
_CN.TRAINER.ADAM_DECAY = 0.0 # ADAM: for adam
_CN.TRAINER.ADAMW_DECAY = 0.1
# step-based warm-up
_CN.TRAINER.WARMUP_TYPE = "linear" # [linear, constant]
_CN.TRAINER.WARMUP_RATIO = 0.0
_CN.TRAINER.WARMUP_STEP = 4800
# learning rate scheduler
# [MultiStepLR, CosineAnnealing, ExponentialLR]
_CN.TRAINER.SCHEDULER = "MultiStepLR"
_CN.TRAINER.SCHEDULER_INTERVAL = "epoch" # [epoch, step]
_CN.TRAINER.MSLR_MILESTONES = [3, 6, 9, 12] # MSLR: MultiStepLR
_CN.TRAINER.MSLR_GAMMA = 0.5
_CN.TRAINER.COSA_TMAX = 30 # COSA: CosineAnnealing
# ELR: ExponentialLR, this value for 'step' interval
_CN.TRAINER.ELR_GAMMA = 0.999992
# plotting related
_CN.TRAINER.ENABLE_PLOTTING = False
_CN.TRAINER.N_VAL_PAIRS_TO_PLOT = 32 # number of val/test paris for plotting
_CN.TRAINER.PLOT_MODE = "evaluation" # ['evaluation', 'confidence']
_CN.TRAINER.PLOT_MATCHES_ALPHA = "dynamic"
# geometric metrics and pose solver
# recommendation: 5e-4 for ScanNet, 1e-4 for MegaDepth (from SuperGlue)
_CN.TRAINER.EPI_ERR_THR = 5e-4
_CN.TRAINER.POSE_GEO_MODEL = "E" # ['E', 'F', 'H']
_CN.TRAINER.POSE_ESTIMATION_METHOD = "RANSAC" # [RANSAC, LO-RANSAC]
_CN.TRAINER.RANSAC_PIXEL_THR = 0.5
_CN.TRAINER.RANSAC_CONF = 0.99999
# data sampler for train_dataloader
_CN.TRAINER.DATA_SAMPLER = (
"scene_balance" # options: ['scene_balance', 'random', 'normal']
)
# 'scene_balance' config
_CN.TRAINER.N_SAMPLES_PER_SUBSET = 200
# whether sample each scene with replacement or not
_CN.TRAINER.SB_SUBSET_SAMPLE_REPLACEMENT = True
# after sampling from scenes, whether shuffle within the epoch or not
_CN.TRAINER.SB_SUBSET_SHUFFLE = True
_CN.TRAINER.SB_REPEAT = 1 # repeat N times for training the sampled data
# 'random' config
_CN.TRAINER.RDM_REPLACEMENT = True
_CN.TRAINER.RDM_NUM_SAMPLES = None
# gradient clipping
_CN.TRAINER.GRADIENT_CLIPPING = 0.5
# reproducibility
# This seed affects the data sampling. With the same seed, the data sampling is promised
# to be the same. When resume training from a checkpoint, it's better to use a different
# seed, otherwise the sampled data will be exactly the same as before resuming, which will
# cause less unique data items sampled during the entire training.
# Use of different seed values might affect the final training result, since not all data items
# are used during training on ScanNet. (60M pairs of images sampled during traing from 230M pairs in total.)
_CN.TRAINER.SEED = 66
def get_cfg_defaults():
"""Get a yacs CfgNode object with default values for my_project."""
# Return a clone so that the defaults will not be altered
# This is for the "local variable" use pattern
return _CN.clone()