vidalnt commited on
Commit
8974c96
·
verified ·
1 Parent(s): 9c1e941

Upload 3 files

Browse files
Files changed (3) hide show
  1. checkpoint_last.pt +3 -0
  2. hydra_train.log +217 -0
  3. hydra_train2.log +0 -0
checkpoint_last.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a96af868b72e07027aa333664743af18fc9160ecbd7c18cc0ff8748e4a92b63
3
+ size 1497630535
hydra_train.log ADDED
@@ -0,0 +1,217 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2026-02-02 23:13:56,489][fairseq.distributed.utils][INFO] - setting CUDA device=1 on rank 1
2
+ [2026-02-02 23:13:57,103][fairseq.distributed.utils][INFO] - distributed init (rank 0): env://
3
+ [2026-02-02 23:13:57,110][fairseq.distributed.utils][INFO] - initialized host a8491112a835 as rank 0
4
+ [2026-02-02 23:13:57,112][fairseq.distributed.utils][INFO] - distributed init (rank 1): env://
5
+ [2026-02-02 23:13:57,118][fairseq.distributed.utils][INFO] - initialized host a8491112a835 as rank 1
6
+ [2026-02-02 23:13:57,510][fairseq_cli.train][INFO] - {'_name': None, 'common': {'_name': None, 'no_progress_bar': False, 'log_interval': 100, 'log_format': 'json', 'log_file': None, 'aim_repo': None, 'aim_run_hash': None, 'tensorboard_logdir': 'tb_asa_finetune', 'wandb_project': None, 'azureml_logging': False, 'seed': 1, 'cpu': False, 'tpu': False, 'bf16': False, 'memory_efficient_bf16': False, 'fp16': True, 'memory_efficient_fp16': False, 'fp16_no_flatten_grads': False, 'fp16_init_scale': 128, 'fp16_scale_window': None, 'fp16_scale_tolerance': 0.0, 'on_cpu_convert_precision': False, 'min_loss_scale': 0.0001, 'threshold_loss_scale': None, 'amp': False, 'amp_batch_retries': 2, 'amp_init_scale': 128, 'amp_scale_window': None, 'user_dir': 'examples/dinosr', 'empty_cache_freq': 0, 'all_gather_list_size': 16384, 'model_parallel_size': 1, 'quantization_config_path': None, 'profile': False, 'reset_logging': False, 'suppress_crashes': False, 'use_plasma_view': False, 'plasma_path': '/tmp/plasma'}, 'common_eval': {'_name': None, 'path': None, 'post_process': None, 'quiet': False, 'model_overrides': '{}', 'results_path': None}, 'distributed_training': {'_name': None, 'distributed_world_size': 2, 'distributed_num_procs': 2, 'distributed_rank': 0, 'distributed_backend': 'nccl', 'distributed_init_method': 'env://', 'distributed_port': 29500, 'device_id': 0, 'distributed_no_spawn': True, 'ddp_backend': 'pytorch_ddp', 'ddp_comm_hook': 'none', 'bucket_cap_mb': 25, 'fix_batches_to_gpus': False, 'find_unused_parameters': True, 'gradient_as_bucket_view': False, 'fast_stat_sync': False, 'heartbeat_timeout': -1, 'broadcast_buffers': False, 'slowmo_momentum': None, 'slowmo_base_algorithm': 'localsgd', 'localsgd_frequency': 3, 'nprocs_per_node': 2, 'pipeline_model_parallel': False, 'pipeline_balance': None, 'pipeline_devices': None, 'pipeline_chunks': 0, 'pipeline_encoder_balance': None, 'pipeline_encoder_devices': None, 'pipeline_decoder_balance': None, 'pipeline_decoder_devices': None, 'pipeline_checkpoint': 'never', 'zero_sharding': 'none', 'fp16': True, 'memory_efficient_fp16': False, 'tpu': False, 'no_reshard_after_forward': False, 'fp32_reduce_scatter': False, 'cpu_offload': False, 'use_sharded_state': False, 'not_fsdp_flatten_parameters': False}, 'dataset': {'_name': None, 'num_workers': 4, 'skip_invalid_size_inputs_valid_test': True, 'max_tokens': 1400000, 'batch_size': None, 'required_batch_size_multiple': 1, 'required_seq_len_multiple': 1, 'dataset_impl': None, 'data_buffer_size': 10, 'train_subset': 'train', 'valid_subset': 'valid', 'combine_valid_subsets': None, 'ignore_unused_valid_subsets': False, 'validate_interval': 5, 'validate_interval_updates': 0, 'validate_after_updates': 0, 'fixed_validation_seed': None, 'disable_validation': True, 'max_tokens_valid': 1400000, 'batch_size_valid': None, 'max_valid_steps': None, 'curriculum': 0, 'gen_subset': 'test', 'num_shards': 1, 'shard_id': 0, 'grouped_shuffling': False, 'update_epoch_batch_itr': False, 'update_ordered_indices_seed': False}, 'optimization': {'_name': None, 'max_epoch': 0, 'max_update': 25000, 'stop_time_hours': 0.0, 'clip_norm': 0.0, 'sentence_avg': False, 'update_freq': [4], 'lr': [5e-05], 'stop_min_lr': -1.0, 'use_bmuf': False, 'skip_remainder_batch': False, 'debug_param_names': False}, 'checkpoint': {'_name': None, 'save_dir': 'checkpoints', 'restore_file': '/kaggle/working/dinosr_base.ckpt', 'continue_once': None, 'finetune_from_model': None, 'reset_dataloader': False, 'reset_lr_scheduler': True, 'reset_meters': False, 'reset_optimizer': True, 'optimizer_overrides': '{}', 'save_interval': 1, 'save_interval_updates': 2500, 'keep_interval_updates': 1, 'keep_interval_updates_pattern': -1, 'keep_last_epochs': -1, 'keep_best_checkpoints': 1, 'no_save': False, 'no_epoch_checkpoints': True, 'no_last_checkpoints': False, 'no_save_optimizer_state': False, 'best_checkpoint_metric': 'loss', 'maximize_best_checkpoint_metric': False, 'patience': -1, 'checkpoint_suffix': '', 'checkpoint_shard_count': 1, 'load_checkpoint_on_all_dp_ranks': True, 'write_checkpoints_asynchronously': False, 'model_parallel_size': 1}, 'bmuf': {'_name': None, 'block_lr': 1.0, 'block_momentum': 0.875, 'global_sync_iter': 50, 'warmup_iterations': 500, 'use_nbm': False, 'average_sync': False, 'distributed_world_size': 2}, 'generation': {'_name': None, 'beam': 5, 'beam_mt': 0, 'nbest': 1, 'max_len_a': 0.0, 'max_len_b': 200, 'max_len_a_mt': 0.0, 'max_len_b_mt': 200, 'min_len': 1, 'match_source_len': False, 'unnormalized': False, 'no_early_stop': False, 'no_beamable_mm': False, 'lenpen': 1.0, 'lenpen_mt': 1.0, 'unkpen': 0.0, 'replace_unk': None, 'sacrebleu': False, 'score_reference': False, 'prefix_size': 0, 'no_repeat_ngram_size': 0, 'sampling': False, 'sampling_topk': -1, 'sampling_topp': -1.0, 'constraints': None, 'temperature': 1.0, 'diverse_beam_groups': -1, 'diverse_beam_strength': 0.5, 'diversity_rate': -1.0, 'print_alignment': None, 'print_step': False, 'lm_path': None, 'lm_weight': 0.0, 'iter_decode_eos_penalty': 0.0, 'iter_decode_max_iter': 10, 'iter_decode_force_max_iter': False, 'iter_decode_with_beam': 1, 'iter_decode_with_external_reranker': False, 'retain_iter_history': False, 'retain_dropout': False, 'retain_dropout_modules': None, 'decoding_format': None, 'no_seed_provided': False, 'eos_token': None}, 'eval_lm': {'_name': None, 'output_word_probs': False, 'output_word_stats': False, 'context_window': 0, 'softmax_batch': 9223372036854775807}, 'interactive': {'_name': None, 'buffer_size': 0, 'input': '-'}, 'model': {'_name': 'dinosr', 'extractor_mode': layer_norm, 'encoder_layers': 12, 'encoder_embed_dim': 768, 'encoder_ffn_embed_dim': 3072, 'encoder_attention_heads': 12, 'activation_fn': gelu, 'layer_type': transformer, 'dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'encoder_layerdrop': 0.05, 'dropout_input': 0.0, 'dropout_features': 0.0, 'final_dim': 0, 'layer_norm_first': False, 'conv_feature_layers': '[(512, 10, 5)] + [(512, 3, 2)] * 4 + [(512,2,2)] + [(512,2,2)]', 'conv_bias': False, 'logit_temp': 0.1, 'quantize_targets': False, 'quantize_input': False, 'same_quantizer': False, 'target_glu': False, 'feature_grad_mult': 0.1, 'quantizer_depth': 1, 'quantizer_factor': 3, 'latent_vars': 320, 'latent_groups': 2, 'latent_dim': 0, 'mask_length': 10, 'mask_prob': 0.8, 'mask_selection': static, 'mask_other': 0.0, 'no_mask_overlap': False, 'mask_min_space': 1, 'require_same_masks': True, 'mask_dropout': 0.0, 'mask_channel_length': 10, 'mask_channel_prob': 0.0, 'mask_channel_before': False, 'mask_channel_selection': static, 'mask_channel_other': 0.0, 'no_mask_channel_overlap': False, 'mask_channel_min_space': 1, 'num_negatives': 100, 'negatives_from_everywhere': False, 'cross_sample_negatives': 0, 'codebook_negatives': 0, 'conv_pos': 95, 'conv_pos_groups': 16, 'pos_conv_depth': 5, 'latent_temp': [2.0, 0.5, 0.999995], 'max_positions': 100000, 'checkpoint_activations': False, 'required_seq_len_multiple': 2, 'crop_seq_to_multiple': 1, 'depthwise_conv_kernel_size': 31, 'attn_type': '', 'pos_enc_type': 'abs', 'fp16': False, 'adp_num': -1, 'adp_dim': 64, 'adp_act_fn': 'relu', 'adp_trf_idx': 'all', 'discrete': True, 'codebook_size': 256, 'normal_init_codebook': False, 'codebook_init_decay': 0.9, 'codebook_end_decay': 0.9, 'codebook_end_decay_step': 0, 'freeze_teacher_step': 200001, 'freeze_pre_enc_modules': True, 'loss_beta': 0.0, 'loss_scale': None, 'average_top_k_layers': 8, 'enable_asa': True, 'layer_norm_target_layer': False, 'instance_norm_target_layer': True, 'instance_norm_targets': False, 'layer_norm_targets': False, 'batch_norm_target_layer': False, 'group_norm_target_layer': False, 'ema_decay': 0.999, 'ema_end_decay': 0.9999, 'ema_anneal_end_step': 15000, 'ema_transformer_only': True, 'ema_layers_only': True, 'max_update': '${optimization.max_update}', 'min_target_var': 0.1, 'min_pred_var': 0.01}, 'task': {'_name': 'audio_pretraining', 'data': '/kaggle/dataset/manifests', 'labels': None, 'multi_corpus_keys': None, 'multi_corpus_sampling_weights': None, 'binarized_dataset': False, 'sample_rate': 16000, 'normalize': True, 'enable_padding': False, 'max_sample_size': 320000, 'min_sample_size': 32000, 'num_batch_buckets': 0, 'tpu': False, 'text_compression_level': none, 'rebuild_batches': True, 'precompute_mask_config': None, 'post_save_script': None, 'subsample': 1.0, 'seed': 1}, 'criterion': {'_name': 'model', 'loss_weights': {}, 'log_keys': ['ema_decay', 'target_ppl', 'pred_ppl', 'codebook_decay', 'asa_I_mu', 'asa_I_sigma', 'asa_Sigma_prime_mu', 'asa_Sigma_prime_sigma'], 'can_sum': True}, 'optimizer': {'_name': 'adam', 'adam_betas': '(0.9,0.98)', 'adam_eps': 1e-06, 'weight_decay': 0.01, 'use_old_adam': False, 'fp16_adam_stats': False, 'tpu': False, 'lr': [5e-05]}, 'lr_scheduler': {'_name': 'tri_stage', 'warmup_steps': 0, 'hold_steps': 0, 'decay_steps': 0, 'phase_ratio': [0.1, 0.4, 0.5], 'init_lr_scale': 0.01, 'final_lr_scale': 0.01, 'max_update': 25000.0, 'lr': [5e-05]}, 'scoring': None, 'bpe': None, 'tokenizer': None, 'ema': {'_name': None, 'store_ema': False, 'ema_decay': 0.9999, 'ema_start_update': 0, 'ema_seed_model': None, 'ema_update_freq': 1, 'ema_fp32': False}, 'job_logging_cfg': {'version': 1, 'formatters': {'simple': {'format': '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'}}, 'handlers': {'console': {'class': 'logging.StreamHandler', 'formatter': 'simple', 'stream': 'ext://sys.stdout'}, 'file': {'class': 'logging.FileHandler', 'formatter': 'simple', 'filename': '/kaggle/working/fairseq/outputs/2026-02-02/23-13-56/hydra_train.log'}}, 'root': {'level': 'INFO', 'handlers': ['console', 'file']}, 'disable_existing_loggers': False}}
7
+ [2026-02-02 23:13:57,585][dinosr.models.dinosr][INFO] - SAVC: Adversarial Style Augmentation enabled.
8
+ [2026-02-02 23:13:59,050][fairseq_cli.train][INFO] - DinosrModel(
9
+ (feature_extractor): ConvFeatureExtractionModel(
10
+ (conv_layers): ModuleList(
11
+ (0): Sequential(
12
+ (0): Conv1d(1, 512, kernel_size=(10,), stride=(5,), bias=False)
13
+ (1): Dropout(p=0.0, inplace=False)
14
+ (2): Sequential(
15
+ (0): TransposeLast()
16
+ (1): Fp32LayerNorm((512,), eps=1e-05, elementwise_affine=True)
17
+ (2): TransposeLast()
18
+ )
19
+ (3): GELU(approximate='none')
20
+ )
21
+ (1-4): 4 x Sequential(
22
+ (0): Conv1d(512, 512, kernel_size=(3,), stride=(2,), bias=False)
23
+ (1): Dropout(p=0.0, inplace=False)
24
+ (2): Sequential(
25
+ (0): TransposeLast()
26
+ (1): Fp32LayerNorm((512,), eps=1e-05, elementwise_affine=True)
27
+ (2): TransposeLast()
28
+ )
29
+ (3): GELU(approximate='none')
30
+ )
31
+ (5-6): 2 x Sequential(
32
+ (0): Conv1d(512, 512, kernel_size=(2,), stride=(2,), bias=False)
33
+ (1): Dropout(p=0.0, inplace=False)
34
+ (2): Sequential(
35
+ (0): TransposeLast()
36
+ (1): Fp32LayerNorm((512,), eps=1e-05, elementwise_affine=True)
37
+ (2): TransposeLast()
38
+ )
39
+ (3): GELU(approximate='none')
40
+ )
41
+ )
42
+ )
43
+ (asa_module): SAVC_ASA(
44
+ (grl): GradientReversal()
45
+ )
46
+ (post_extract_proj): Linear(in_features=512, out_features=768, bias=True)
47
+ (dropout_input): Dropout(p=0.0, inplace=False)
48
+ (dropout_features): Dropout(p=0.0, inplace=False)
49
+ (encoder): TransformerEncoder(
50
+ (pos_conv): Sequential(
51
+ (0): Sequential(
52
+ (0): Conv1d(768, 768, kernel_size=(19,), stride=(1,), padding=(9,), groups=16)
53
+ (1): SamePad()
54
+ (2): TransposeLast()
55
+ (3): LayerNorm((768,), eps=1e-05, elementwise_affine=False)
56
+ (4): TransposeLast()
57
+ (5): GELU(approximate='none')
58
+ )
59
+ (1): Sequential(
60
+ (0): Conv1d(768, 768, kernel_size=(19,), stride=(1,), padding=(9,), groups=16)
61
+ (1): SamePad()
62
+ (2): TransposeLast()
63
+ (3): LayerNorm((768,), eps=1e-05, elementwise_affine=False)
64
+ (4): TransposeLast()
65
+ (5): GELU(approximate='none')
66
+ )
67
+ (2): Sequential(
68
+ (0): Conv1d(768, 768, kernel_size=(19,), stride=(1,), padding=(9,), groups=16)
69
+ (1): SamePad()
70
+ (2): TransposeLast()
71
+ (3): LayerNorm((768,), eps=1e-05, elementwise_affine=False)
72
+ (4): TransposeLast()
73
+ (5): GELU(approximate='none')
74
+ )
75
+ (3): Sequential(
76
+ (0): Conv1d(768, 768, kernel_size=(19,), stride=(1,), padding=(9,), groups=16)
77
+ (1): SamePad()
78
+ (2): TransposeLast()
79
+ (3): LayerNorm((768,), eps=1e-05, elementwise_affine=False)
80
+ (4): TransposeLast()
81
+ (5): GELU(approximate='none')
82
+ )
83
+ (4): Sequential(
84
+ (0): Conv1d(768, 768, kernel_size=(19,), stride=(1,), padding=(9,), groups=16)
85
+ (1): SamePad()
86
+ (2): TransposeLast()
87
+ (3): LayerNorm((768,), eps=1e-05, elementwise_affine=False)
88
+ (4): TransposeLast()
89
+ (5): GELU(approximate='none')
90
+ )
91
+ )
92
+ (layers): ModuleList(
93
+ (0-11): 12 x TransformerSentenceEncoderLayer(
94
+ (self_attn): MultiheadAttention(
95
+ (dropout_module): FairseqDropout()
96
+ (k_proj): Linear(in_features=768, out_features=768, bias=True)
97
+ (v_proj): Linear(in_features=768, out_features=768, bias=True)
98
+ (q_proj): Linear(in_features=768, out_features=768, bias=True)
99
+ (out_proj): Linear(in_features=768, out_features=768, bias=True)
100
+ )
101
+ (dropout1): Dropout(p=0.1, inplace=False)
102
+ (dropout2): Dropout(p=0.0, inplace=False)
103
+ (dropout3): Dropout(p=0.1, inplace=False)
104
+ (self_attn_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
105
+ (fc1): Linear(in_features=768, out_features=3072, bias=True)
106
+ (fc2): Linear(in_features=3072, out_features=768, bias=True)
107
+ (final_layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
108
+ )
109
+ )
110
+ (layer_norm): LayerNorm((768,), eps=1e-05, elementwise_affine=True)
111
+ )
112
+ (layer_norm): LayerNorm((512,), eps=1e-05, elementwise_affine=True)
113
+ (heads): ModuleList(
114
+ (0-7): 8 x Linear(in_features=768, out_features=256, bias=True)
115
+ )
116
+ )
117
+ [2026-02-02 23:13:59,053][fairseq_cli.train][INFO] - task: AudioPretrainingTask
118
+ [2026-02-02 23:13:59,053][fairseq_cli.train][INFO] - model: DinosrModel
119
+ [2026-02-02 23:13:59,053][fairseq_cli.train][INFO] - criterion: ModelCriterion
120
+ [2026-02-02 23:13:59,054][fairseq_cli.train][INFO] - num. shared model params: 94,740,224 (num. trained: 94,740,224)
121
+ [2026-02-02 23:13:59,055][fairseq_cli.train][INFO] - num. expert model params: 0 (num. trained: 0)
122
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.1.0.bias
123
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.2.0.bias
124
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.3.0.bias
125
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.4.0.bias
126
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.5.0.bias
127
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- feature_extractor.conv_layers.6.0.bias
128
+ [2026-02-02 23:13:59,171][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.0.3.weight
129
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.0.3.bias
130
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.1.3.weight
131
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.1.3.bias
132
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.2.3.weight
133
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.2.3.bias
134
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.3.3.weight
135
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.3.3.bias
136
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.4.3.weight
137
+ [2026-02-02 23:13:59,172][fairseq.trainer][INFO] - detected shared parameter: feature_extractor.conv_layers.0.0.bias <- encoder.pos_conv.4.3.bias
138
+ [2026-02-02 23:13:59,240][fairseq.utils][INFO] - ***********************CUDA enviroments for all 2 workers***********************
139
+ [2026-02-02 23:13:59,241][fairseq.utils][INFO] - rank 0: capabilities = 7.5 ; total memory = 14.563 GB ; name = Tesla T4
140
+ [2026-02-02 23:13:59,241][fairseq.utils][INFO] - rank 1: capabilities = 7.5 ; total memory = 14.563 GB ; name = Tesla T4
141
+ [2026-02-02 23:13:59,241][fairseq.utils][INFO] - ***********************CUDA enviroments for all 2 workers***********************
142
+ [2026-02-02 23:13:59,241][fairseq_cli.train][INFO] - training on 2 devices (GPUs/TPUs)
143
+ [2026-02-02 23:13:59,241][fairseq_cli.train][INFO] - max tokens per device = 1400000 and max sentences per device = None
144
+ [2026-02-02 23:13:59,242][fairseq.trainer][INFO] - Preparing to load checkpoint /kaggle/working/dinosr_base.ckpt
145
+ [2026-02-02 23:14:01,494][dinosr.models.dinosr][INFO] - making ema teacher
146
+ [2026-02-02 23:14:01,836][fairseq.trainer][INFO] - Loaded checkpoint /kaggle/working/dinosr_base.ckpt (epoch 428 @ 0 updates)
147
+ [2026-02-02 23:14:01,884][fairseq.trainer][INFO] - loading train data for epoch 428
148
+ [2026-02-02 23:14:01,906][fairseq.data.audio.raw_audio_dataset][INFO] - loaded 27124, skipped 22 samples
149
+ [2026-02-02 23:14:01,914][fairseq.tasks.fairseq_task][INFO] - can_reuse_epoch_itr = True
150
+ [2026-02-02 23:14:01,915][fairseq.tasks.fairseq_task][INFO] - reuse_dataloader = True
151
+ [2026-02-02 23:14:01,915][fairseq.tasks.fairseq_task][INFO] - rebuild_batches = True
152
+ [2026-02-02 23:14:01,915][fairseq.tasks.fairseq_task][INFO] - batches will be rebuilt for each epoch
153
+ [2026-02-02 23:14:01,915][fairseq.tasks.fairseq_task][INFO] - creating new batches for epoch 428
154
+ [2026-02-02 23:14:01,969][fairseq.data.iterators][INFO] - grouped total_num_itrs = 539
155
+ [2026-02-02 23:14:01,972][fairseq.trainer][INFO] - begin training epoch 428
156
+ [2026-02-02 23:14:01,972][fairseq_cli.train][INFO] - Start iterating over samples
157
+ [2026-02-02 23:14:06,867][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 64.0
158
+ [2026-02-02 23:14:09,870][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 32.0
159
+ [2026-02-02 23:14:12,814][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 16.0
160
+ [2026-02-02 23:14:15,725][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 8.0
161
+ [2026-02-02 23:14:18,668][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 4.0
162
+ [2026-02-02 23:14:21,533][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 2.0
163
+ [2026-02-02 23:14:24,443][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 1.0
164
+ [2026-02-02 23:14:27,308][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.5
165
+ [2026-02-02 23:14:33,114][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.25
166
+ [2026-02-02 23:14:35,985][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.125
167
+ [2026-02-02 23:14:53,388][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.0625
168
+ [2026-02-02 23:15:17,206][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.03125
169
+ [2026-02-02 23:19:57,349][train_inner][INFO] - {"epoch": 428, "update": 427.301, "loss": "5.304", "ntokens": "17008.6", "nsentences": "50.68", "ema_decay": "999.003", "target_ppl": "136.565", "pred_ppl": "107.323", "codebook_decay": "0.9", "sample_size": "17008.6", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "4784.1", "ups": "0.28", "wpb": "17008.6", "bsz": "50.7", "num_updates": "100", "lr": "2.48e-06", "gnorm": "11.353", "loss_scale": "0.0312", "train_wall": "354", "gb_free": "8", "wall": "0"}
170
+ [2026-02-02 23:25:25,450][train_inner][INFO] - {"epoch": 428, "update": 427.486, "loss": "4.89", "ntokens": "16939.6", "nsentences": "49.89", "ema_decay": "999.01", "target_ppl": "135.689", "pred_ppl": "147.797", "codebook_decay": "0.9", "sample_size": "16939.6", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "5163", "ups": "0.3", "wpb": "16939.6", "bsz": "49.9", "num_updates": "200", "lr": "4.46e-06", "gnorm": "2.632", "loss_scale": "0.0312", "train_wall": "327", "gb_free": "8.2", "wall": "0"}
171
+ [2026-02-02 23:25:45,345][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.015625
172
+ [2026-02-02 23:30:56,900][train_inner][INFO] - {"epoch": 428, "update": 427.673, "loss": "4.766", "ntokens": "16942.9", "nsentences": "51.14", "ema_decay": "999.017", "target_ppl": "134.889", "pred_ppl": "150.375", "codebook_decay": "0.9", "sample_size": "16942.9", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "5111.8", "ups": "0.3", "wpb": "16942.9", "bsz": "51.1", "num_updates": "300", "lr": "6.44e-06", "gnorm": "2.172", "loss_scale": "0.0156", "train_wall": "330", "gb_free": "7.8", "wall": "0"}
173
+ [2026-02-02 23:36:23,526][train_inner][INFO] - {"epoch": 428, "update": 427.859, "loss": "4.642", "ntokens": "16985.2", "nsentences": "49.97", "ema_decay": "999.023", "target_ppl": "133.678", "pred_ppl": "149.727", "codebook_decay": "0.9", "sample_size": "16985.2", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "5200.3", "ups": "0.31", "wpb": "16985.2", "bsz": "50", "num_updates": "400", "lr": "8.42e-06", "gnorm": "1.829", "loss_scale": "0.0156", "train_wall": "325", "gb_free": "8.3", "wall": "0"}
174
+ [2026-02-02 23:40:31,854][fairseq.checkpoint_utils][INFO] - Preparing to save checkpoint for epoch 428 @ 476 updates
175
+ [2026-02-02 23:40:31,855][fairseq.trainer][INFO] - Saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
176
+ [2026-02-02 23:40:34,649][fairseq.trainer][INFO] - Finished saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
177
+ [2026-02-02 23:40:34,650][fairseq.checkpoint_utils][INFO] - Saved checkpoint checkpoints/checkpoint_last.pt (epoch 428 @ 476 updates, score None) (writing took 2.7960509399999864 seconds)
178
+ [2026-02-02 23:40:34,652][fairseq_cli.train][INFO] - end of epoch 428 (average epoch stats below)
179
+ [2026-02-02 23:40:34,655][train][INFO] - {"epoch": 428, "train_loss": "2.242", "train_ntokens": "39807.8", "train_nsentences": "124.018", "train_ema_decay": "999.304", "train_target_ppl": "139.531", "train_pred_ppl": "143.484", "train_codebook_decay": "0.929", "train_sample_size": "16977.8", "train_asa_I_mu": "1", "train_asa_I_sigma": "1", "train_asa_Sigma_prime_mu": "0.966", "train_asa_Sigma_prime_sigma": "0.974", "train_wps": "14442.6", "train_ups": "0.36", "train_wpb": "39807.8", "train_bsz": "124", "train_num_updates": "476", "train_lr": "9.9248e-06", "train_gnorm": "3.135", "train_loss_scale": "0.0156", "train_train_wall": "1838", "train_gb_free": "8.1", "train_wall": "0"}
180
+ [2026-02-02 23:40:34,656][fairseq.tasks.fairseq_task][INFO] - can_reuse_epoch_itr = True
181
+ [2026-02-02 23:40:34,694][fairseq.data.iterators][INFO] - grouped total_num_itrs = 539
182
+ [2026-02-02 23:40:34,697][fairseq.trainer][INFO] - begin training epoch 428
183
+ [2026-02-02 23:40:34,697][fairseq_cli.train][INFO] - Start iterating over samples
184
+ [2026-02-02 23:41:54,466][train_inner][INFO] - {"epoch": 428, "update": 427.045, "loss": "4.507", "ntokens": "17042.7", "nsentences": "50.48", "ema_decay": "999.03", "target_ppl": "133.149", "pred_ppl": "148.784", "codebook_decay": "0.9", "sample_size": "17042.7", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "5149.8", "ups": "0.3", "wpb": "17042.7", "bsz": "50.5", "num_updates": "500", "lr": "1.04e-05", "gnorm": "1.391", "loss_scale": "0.0156", "train_wall": "327", "gb_free": "8.2", "wall": "0"}
185
+ [2026-02-02 23:47:20,798][train_inner][INFO] - {"epoch": 428, "update": 427.23, "loss": "4.376", "ntokens": "16964.8", "nsentences": "50.33", "ema_decay": "999.037", "target_ppl": "132.864", "pred_ppl": "148.479", "codebook_decay": "0.9", "sample_size": "16964.8", "asa_I_mu": "1", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.974", "wps": "5198.7", "ups": "0.31", "wpb": "16964.8", "bsz": "50.3", "num_updates": "600", "lr": "1.238e-05", "gnorm": "1.965", "loss_scale": "0.0156", "train_wall": "325", "gb_free": "8.2", "wall": "0"}
186
+ [2026-02-02 23:52:48,491][train_inner][INFO] - {"epoch": 428, "update": 427.416, "loss": "4.266", "ntokens": "16981.7", "nsentences": "50.23", "ema_decay": "999.043", "target_ppl": "132.45", "pred_ppl": "147.933", "codebook_decay": "0.9", "sample_size": "16981.7", "asa_I_mu": "1.001", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.966", "asa_Sigma_prime_sigma": "0.973", "wps": "5182.2", "ups": "0.31", "wpb": "16981.7", "bsz": "50.2", "num_updates": "700", "lr": "1.436e-05", "gnorm": "1.478", "loss_scale": "0.0156", "train_wall": "327", "gb_free": "7.9", "wall": "0"}
187
+ [2026-02-02 23:58:17,112][train_inner][INFO] - {"epoch": 428, "update": 427.601, "loss": "4.146", "ntokens": "16925.1", "nsentences": "50.96", "ema_decay": "999.05", "target_ppl": "131.881", "pred_ppl": "147.057", "codebook_decay": "0.9", "sample_size": "16925.1", "asa_I_mu": "1.001", "asa_I_sigma": "1", "asa_Sigma_prime_mu": "0.965", "asa_Sigma_prime_sigma": "0.973", "wps": "5150.4", "ups": "0.3", "wpb": "16925.1", "bsz": "51", "num_updates": "800", "lr": "1.634e-05", "gnorm": "1.261", "loss_scale": "0.0156", "train_wall": "328", "gb_free": "8", "wall": "0"}
188
+ [2026-02-02 23:59:32,435][fairseq.trainer][INFO] - NOTE: gradient overflow detected, ignoring gradient, setting loss scale to: 0.0078125
189
+ [2026-02-03 00:03:48,216][train_inner][INFO] - {"epoch": 428, "update": 427.788, "loss": "4.024", "ntokens": "16946.3", "nsentences": "50.56", "ema_decay": "999.057", "target_ppl": "131.805", "pred_ppl": "146.525", "codebook_decay": "0.9", "sample_size": "16946.3", "asa_I_mu": "1.001", "asa_I_sigma": "1.001", "asa_Sigma_prime_mu": "0.965", "asa_Sigma_prime_sigma": "0.972", "wps": "5118.2", "ups": "0.3", "wpb": "16946.3", "bsz": "50.6", "num_updates": "900", "lr": "1.832e-05", "gnorm": "1.625", "loss_scale": "0.0078", "train_wall": "330", "gb_free": "8", "wall": "0"}
190
+ [2026-02-03 00:09:13,939][train_inner][INFO] - {"epoch": 428, "update": 427.974, "loss": "3.898", "ntokens": "16979.8", "nsentences": "49.99", "ema_decay": "999.063", "target_ppl": "131.923", "pred_ppl": "146.114", "codebook_decay": "0.9", "sample_size": "16979.8", "asa_I_mu": "1.001", "asa_I_sigma": "1.001", "asa_Sigma_prime_mu": "0.964", "asa_Sigma_prime_sigma": "0.971", "wps": "5213.1", "ups": "0.31", "wpb": "16979.8", "bsz": "50", "num_updates": "1000", "lr": "2.03e-05", "gnorm": "1.406", "loss_scale": "0.0078", "train_wall": "325", "gb_free": "8", "wall": "0"}
191
+ [2026-02-03 00:09:57,404][fairseq.checkpoint_utils][INFO] - Preparing to save checkpoint for epoch 428 @ 1014 updates
192
+ [2026-02-03 00:09:57,405][fairseq.trainer][INFO] - Saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
193
+ [2026-02-03 00:10:01,186][fairseq.trainer][INFO] - Finished saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
194
+ [2026-02-03 00:10:01,187][fairseq.checkpoint_utils][INFO] - Saved checkpoint checkpoints/checkpoint_last.pt (epoch 428 @ 1014 updates, score None) (writing took 3.782398663999629 seconds)
195
+ [2026-02-03 00:10:01,188][fairseq_cli.train][INFO] - end of epoch 428 (average epoch stats below)
196
+ [2026-02-03 00:10:01,190][train][INFO] - {"epoch": 428, "train_loss": "4.147", "train_ntokens": "16951.5", "train_nsentences": "50.3364", "train_ema_decay": "999.05", "train_target_ppl": "132.223", "train_pred_ppl": "147.242", "train_codebook_decay": "0.9", "train_sample_size": "16951.5", "train_asa_I_mu": "1.001", "train_asa_I_sigma": "1", "train_asa_Sigma_prime_mu": "0.965", "train_asa_Sigma_prime_sigma": "0.973", "train_wps": "5162.6", "train_ups": "0.3", "train_wpb": "16951.5", "train_bsz": "50.3", "train_num_updates": "1014", "train_lr": "2.05772e-05", "train_gnorm": "1.539", "train_loss_scale": "0.0078", "train_train_wall": "1757", "train_gb_free": "8.3", "train_wall": "0"}
197
+ [2026-02-03 00:10:01,191][fairseq.tasks.fairseq_task][INFO] - can_reuse_epoch_itr = True
198
+ [2026-02-03 00:10:01,225][fairseq.tasks.fairseq_task][INFO] - creating new batches for epoch 429
199
+ [2026-02-03 00:10:01,249][fairseq.data.iterators][INFO] - grouped total_num_itrs = 539
200
+ [2026-02-03 00:10:01,252][fairseq.trainer][INFO] - begin training epoch 429
201
+ [2026-02-03 00:10:01,252][fairseq_cli.train][INFO] - Start iterating over samples
202
+ [2026-02-03 00:14:42,597][train_inner][INFO] - {"epoch": 429, "update": 428.16, "loss": "3.755", "ntokens": "16890", "nsentences": "48", "ema_decay": "999.07", "target_ppl": "131.975", "pred_ppl": "145.727", "codebook_decay": "0.9", "sample_size": "16890", "asa_I_mu": "1.002", "asa_I_sigma": "1.001", "asa_Sigma_prime_mu": "0.963", "asa_Sigma_prime_sigma": "0.971", "wps": "5139.2", "ups": "0.3", "wpb": "16890", "bsz": "48", "num_updates": "1100", "lr": "2.228e-05", "gnorm": "1.622", "loss_scale": "0.0078", "train_wall": "324", "gb_free": "8.1", "wall": "0"}
203
+ [2026-02-03 00:20:08,394][train_inner][INFO] - {"epoch": 429, "update": 428.345, "loss": "3.629", "ntokens": "16926.7", "nsentences": "51.43", "ema_decay": "999.077", "target_ppl": "131.592", "pred_ppl": "145.048", "codebook_decay": "0.9", "sample_size": "16926.7", "asa_I_mu": "1.002", "asa_I_sigma": "1.002", "asa_Sigma_prime_mu": "0.962", "asa_Sigma_prime_sigma": "0.97", "wps": "5195.6", "ups": "0.31", "wpb": "16926.7", "bsz": "51.4", "num_updates": "1200", "lr": "2.426e-05", "gnorm": "1.542", "loss_scale": "0.0078", "train_wall": "325", "gb_free": "8.1", "wall": "0"}
204
+ [2026-02-03 00:25:36,806][train_inner][INFO] - {"epoch": 429, "update": 428.531, "loss": "3.465", "ntokens": "16967.5", "nsentences": "50.38", "ema_decay": "999.083", "target_ppl": "132.038", "pred_ppl": "144.868", "codebook_decay": "0.9", "sample_size": "16967.5", "asa_I_mu": "1.003", "asa_I_sigma": "1.002", "asa_Sigma_prime_mu": "0.961", "asa_Sigma_prime_sigma": "0.969", "wps": "5166.6", "ups": "0.3", "wpb": "16967.5", "bsz": "50.4", "num_updates": "1300", "lr": "2.624e-05", "gnorm": "1.577", "loss_scale": "0.0078", "train_wall": "327", "gb_free": "8", "wall": "0"}
205
+ [2026-02-03 00:31:04,453][train_inner][INFO] - {"epoch": 429, "update": 428.716, "loss": "3.341", "ntokens": "16964.5", "nsentences": "50.09", "ema_decay": "999.09", "target_ppl": "131.531", "pred_ppl": "144.034", "codebook_decay": "0.9", "sample_size": "16964.5", "asa_I_mu": "1.003", "asa_I_sigma": "1.003", "asa_Sigma_prime_mu": "0.96", "asa_Sigma_prime_sigma": "0.969", "wps": "5177.7", "ups": "0.31", "wpb": "16964.5", "bsz": "50.1", "num_updates": "1400", "lr": "2.822e-05", "gnorm": "1.746", "loss_scale": "0.0078", "train_wall": "327", "gb_free": "8.1", "wall": "0"}
206
+ [2026-02-03 00:36:30,253][train_inner][INFO] - {"epoch": 429, "update": 428.902, "loss": "3.224", "ntokens": "16992.8", "nsentences": "51.68", "ema_decay": "999.097", "target_ppl": "131.978", "pred_ppl": "144.071", "codebook_decay": "0.9", "sample_size": "16992.8", "asa_I_mu": "1.004", "asa_I_sigma": "1.004", "asa_Sigma_prime_mu": "0.959", "asa_Sigma_prime_sigma": "0.968", "wps": "5215.7", "ups": "0.31", "wpb": "16992.8", "bsz": "51.7", "num_updates": "1500", "lr": "3.02e-05", "gnorm": "1.647", "loss_scale": "0.0078", "train_wall": "325", "gb_free": "7.9", "wall": "0"}
207
+ [2026-02-03 00:39:22,389][fairseq.checkpoint_utils][INFO] - Preparing to save checkpoint for epoch 429 @ 1553 updates
208
+ [2026-02-03 00:39:22,390][fairseq.trainer][INFO] - Saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
209
+ [2026-02-03 00:39:25,941][fairseq.trainer][INFO] - Finished saving checkpoint to /kaggle/working/fairseq/outputs/2026-02-02/23-13-56/checkpoints/checkpoint_last.pt
210
+ [2026-02-03 00:39:25,942][fairseq.checkpoint_utils][INFO] - Saved checkpoint checkpoints/checkpoint_last.pt (epoch 429 @ 1553 updates, score None) (writing took 3.553410245000123 seconds)
211
+ [2026-02-03 00:39:25,943][fairseq_cli.train][INFO] - end of epoch 429 (average epoch stats below)
212
+ [2026-02-03 00:39:25,951][train][INFO] - {"epoch": 429, "train_loss": "3.437", "train_ntokens": "16952.7", "train_nsentences": "50.3228", "train_ema_decay": "999.086", "train_target_ppl": "131.832", "train_pred_ppl": "144.62", "train_codebook_decay": "0.9", "train_sample_size": "16952.7", "train_asa_I_mu": "1.003", "train_asa_I_sigma": "1.003", "train_asa_Sigma_prime_mu": "0.961", "train_asa_Sigma_prime_sigma": "0.969", "train_wps": "5177.8", "train_ups": "0.31", "train_wpb": "16952.7", "train_bsz": "50.3", "train_num_updates": "1553", "train_lr": "3.12494e-05", "train_gnorm": "1.636", "train_loss_scale": "0.0078", "train_train_wall": "1755", "train_gb_free": "8", "train_wall": "0"}
213
+ [2026-02-03 00:39:25,952][fairseq.tasks.fairseq_task][INFO] - can_reuse_epoch_itr = True
214
+ [2026-02-03 00:39:25,984][fairseq.tasks.fairseq_task][INFO] - creating new batches for epoch 430
215
+ [2026-02-03 00:39:26,007][fairseq.data.iterators][INFO] - grouped total_num_itrs = 539
216
+ [2026-02-03 00:39:26,010][fairseq.trainer][INFO] - begin training epoch 430
217
+ [2026-02-03 00:39:26,010][fairseq_cli.train][INFO] - Start iterating over samples
hydra_train2.log ADDED
The diff for this file is too large to render. See raw diff