| |
|
|
| common: |
| fp16: true |
| log_format: json |
| log_interval: 200 |
| tensorboard_logdir: tb |
| min_loss_scale: 1e-6 |
| fp16_no_flatten_grads: false |
| user_dir: ${env:PWD}/examples/data2vec |
|
|
| checkpoint: |
| save_interval: 1 |
| save_interval_updates: 25000 |
| keep_interval_updates: 1 |
| no_epoch_checkpoints: true |
|
|
| task: |
| _name: audio_pretraining |
| data: /private/home/abaevski/data/librispeech/full |
| max_sample_size: 320000 |
| min_sample_size: 32000 |
| normalize: true |
| precompute_mask_config: {} |
|
|
| dataset: |
| num_workers: 6 |
| max_tokens: 1000000 |
| skip_invalid_size_inputs_valid_test: true |
| validate_interval: 5 |
| required_batch_size_multiple: 1 |
| disable_validation: true |
|
|
| distributed_training: |
| distributed_world_size: 8 |
| ddp_backend: legacy_ddp |
|
|
| criterion: |
| _name: model |
| log_keys: |
| - ema_decay |
| - target_var |
| - pred_var |
| - model_norm |
| - ema_norm |
| - masked_pct |
|
|
| optimization: |
| max_update: 400000 |
| lr: [0.00075] |
| debug_param_names: true |
|
|
| optimizer: |
| _name: adam |
| adam_betas: [ 0.9,0.98 ] |
| adam_eps: 1e-06 |
| weight_decay: 0.01 |
|
|
| lr_scheduler: |
| _name: cosine |
| warmup_updates: 8000 |
|
|
| model: |
| _name: data2vec_multi |
|
|
| loss_beta: 0 |
| loss_scale: null |
|
|
| depth: 12 |
| embed_dim: 768 |
| clone_batch: 8 |
|
|
| ema_decay: 0.999 |
| ema_end_decay: 0.99999 |
| ema_anneal_end_step: 75000 |
| ema_encoder_only: false |
|
|
| average_top_k_layers: 8 |
| instance_norm_target_layer: true |
| layer_norm_target_layer: false |
| layer_norm_targets: false |
|
|
| layerdrop: 0.05 |
| norm_eps: 1e-5 |
|
|
| supported_modality: AUDIO |
|
|
| modalities: |
| audio: |
| feature_encoder_spec: '[(512, 10, 5)] + [(512, 3, 2)] * 4 + [(512,2,2)] + [(512,2,2)]' |
| conv_pos_depth: 5 |
| conv_pos_width: 95 |
| conv_pos_groups: 16 |
| prenet_depth: 0 |
| mask_prob: 0.5 |
| mask_prob_adjust: 0.05 |
| inverse_mask: false |
| mask_length: 5 |
| mask_noise_std: 0.01 |
| mask_dropout: 0 |
| add_masks: false |
| ema_local_encoder: false |
| use_alibi_encoder: true |
| prenet_layerdrop: 0.05 |
| prenet_dropout: 0.1 |
| learned_alibi_scale: true |
| learned_alibi_scale_per_head: true |
| decoder: |
| input_dropout: 0.1 |
| decoder_dim: 384 |
| decoder_groups: 16 |
| decoder_kernel: 7 |
| decoder_layers: 4 |
|
|