StreamSpeech / fairseq /examples /data2vec /config /vision /pretraining /base_imagenet_d2v1.yaml
fasdfsa's picture
init
901e06a
# @package _group_
common:
fp16: true
log_format: json
log_interval: 200
tensorboard_logdir: tb
checkpoint:
save_interval: 5
save_interval_updates: 25000
keep_interval_updates: 1
no_epoch_checkpoints: true
task:
_name: image_pretraining
data: /datasets01/imagenet_full_size/061417
dataset:
num_workers: 6
batch_size: 128
skip_invalid_size_inputs_valid_test: true
required_batch_size_multiple: 2
disable_validation: true
distributed_training:
distributed_world_size: 16
ddp_backend: legacy_ddp
criterion:
_name: model
log_keys:
- ema_decay
- target_var
- pred_var
optimization:
max_update: 375300 #300*1251
lr: [0.0005]
clip_norm: 3.0
optimizer:
_name: adam
adam_betas: (0.9,0.999)
adam_eps: 1e-08
weight_decay: 0.05
lr_scheduler:
_name: cosine
warmup_updates: 12510 # it should be 10 epochs
model:
_name: data2vec_vision
attention_dropout: 0.05
ema_decay: 0.999
ema_end_decay: 0.9998
layer_norm_targets: True
average_top_k_layers: 6
loss_beta: 2.0
drop_path: 0.25