File size: 1,597 Bytes
1633fcc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
#### access DINOv2
export CUDA_VISIBLE_DEVICES=6;
python fast_vis_settings_all.py \
--batch-size 64 --warmup-epochs 5 --epochs 300 \
--data-set IMNET --data-path '/data1/datasets/imagenet_fold' \
--teacher-model vit_large --target_model vit_base --model models_proteus_dinov2 \
--patch_size 14 --mask_probability 0.5 --mask_ratio 0.5 --mask_first_n \
--lambda_token 1.0 --lambda_fea 1.0 --lambda_patch 1.0 \
--finetune "/data0/qiyp/Proteus-pytorch/pretrain/log/DINOv2_training/checkpoint0160.pth" \
--output_dir log/DINOv2_training;
#### access SynCLR
# python -m torch.distributed.launch --nproc_per_node=4 --use_env main.py \
# --batch-size 128 --warmup-epochs 5 --epochs 300 \
# --data-set IMNET --data-path imagenet_path \
# --teacher-model vit_large --target_model vit_base --model models_proteus_synclr \
# --teacher-path pretrained_synclr_path \
# --patch_size 14 --mask_probability 0.5 --mask_ratio 0.5 --mask_first_n \
# --lambda_token 1.0 --lambda_fea 1.0 --lambda_patch 1.0 \
# --output_dir log/SynCLR_training;
#### access CLIP
# python -m torch.distributed.launch --nproc_per_node=4 --use_env main.py \
# --batch-size 128 --warmup-epochs 5 --epochs 300 \
# --data-set IMNET --data-path imagenet_path \
# --teacher-model vit_large --target_model vit_base --model models_proteus_clip \
# --teacher-path pretrained_clip_path \
# --patch_size 14 --mask_probability 0.5 --mask_ratio 0.5 --mask_first_n \
# --lambda_token 1.0 --lambda_fea 0.0 --lambda_patch 0.0 \
# --output_dir log/CLIP_training; |