Lingwei Meng commited on
Commit
c3e45ef
·
1 Parent(s): e61736d

Add model files

Browse files
Files changed (11) hide show
  1. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/config.yaml +274 -0
  2. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/hydra.yaml +164 -0
  3. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/overrides.yaml +5 -0
  4. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/hydra_train.log +77 -0
  5. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/checkpoint1.pt +3 -0
  6. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/checkpoint2.pt +3 -0
  7. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/multirun.yaml +438 -0
  8. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train.log +0 -0
  9. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train/events.out.tfevents.1738168057.node-0.6519.1 +3 -0
  10. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train_inner/events.out.tfevents.1737792645.node-0.6570.0 +3 -0
  11. lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train_inner/events.out.tfevents.1737992273.node-0.6519.0 +3 -0
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/config.yaml ADDED
@@ -0,0 +1,274 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _name: null
2
+ common:
3
+ _name: null
4
+ no_progress_bar: false
5
+ log_interval: 100
6
+ log_format: json
7
+ log_file: null
8
+ aim_repo: null
9
+ aim_run_hash: null
10
+ tensorboard_logdir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
11
+ wandb_project: null
12
+ azureml_logging: false
13
+ seed: 100501
14
+ cpu: false
15
+ tpu: false
16
+ bf16: false
17
+ memory_efficient_bf16: false
18
+ fp16: true
19
+ memory_efficient_fp16: false
20
+ fp16_no_flatten_grads: false
21
+ fp16_init_scale: 128
22
+ fp16_scale_window: null
23
+ fp16_scale_tolerance: 0.02
24
+ on_cpu_convert_precision: false
25
+ min_loss_scale: 0.0001
26
+ threshold_loss_scale: null
27
+ amp: false
28
+ amp_batch_retries: 2
29
+ amp_init_scale: 128
30
+ amp_scale_window: null
31
+ user_dir: /home/aiscuser/code/speechllm/examples/speechllm/
32
+ empty_cache_freq: 0
33
+ all_gather_list_size: 16384
34
+ model_parallel_size: 1
35
+ quantization_config_path: null
36
+ profile: false
37
+ reset_logging: false
38
+ suppress_crashes: false
39
+ use_plasma_view: false
40
+ plasma_path: /tmp/plasma
41
+ common_eval:
42
+ _name: null
43
+ path: null
44
+ post_process: null
45
+ quiet: false
46
+ model_overrides: '{}'
47
+ results_path: null
48
+ distributed_training:
49
+ _name: null
50
+ distributed_world_size: 16
51
+ distributed_num_procs: 8
52
+ distributed_rank: 0
53
+ distributed_backend: nccl
54
+ distributed_init_method: null
55
+ distributed_port: -1
56
+ device_id: 0
57
+ distributed_no_spawn: false
58
+ ddp_backend: pytorch_ddp
59
+ ddp_comm_hook: none
60
+ bucket_cap_mb: 25
61
+ fix_batches_to_gpus: false
62
+ find_unused_parameters: true
63
+ gradient_as_bucket_view: false
64
+ fast_stat_sync: false
65
+ heartbeat_timeout: -1
66
+ broadcast_buffers: false
67
+ slowmo_momentum: null
68
+ slowmo_base_algorithm: localsgd
69
+ localsgd_frequency: 3
70
+ nprocs_per_node: 8
71
+ pipeline_model_parallel: false
72
+ pipeline_balance: null
73
+ pipeline_devices: null
74
+ pipeline_chunks: 0
75
+ pipeline_encoder_balance: null
76
+ pipeline_encoder_devices: null
77
+ pipeline_decoder_balance: null
78
+ pipeline_decoder_devices: null
79
+ pipeline_checkpoint: never
80
+ zero_sharding: none
81
+ fp16: ${common.fp16}
82
+ memory_efficient_fp16: ${common.memory_efficient_fp16}
83
+ tpu: ${common.tpu}
84
+ no_reshard_after_forward: false
85
+ fp32_reduce_scatter: false
86
+ cpu_offload: false
87
+ use_sharded_state: false
88
+ not_fsdp_flatten_parameters: false
89
+ dataset:
90
+ _name: null
91
+ num_workers: 0
92
+ skip_invalid_size_inputs_valid_test: true
93
+ max_tokens: 960000
94
+ batch_size: null
95
+ required_batch_size_multiple: 1
96
+ required_seq_len_multiple: 1
97
+ dataset_impl: null
98
+ data_buffer_size: 10
99
+ train_subset: libri2mix_train_targetASR,libri3mix_train_targetASR,librispeech2mix_train_targetASR,librispeech3mix_train_targetASR,libri2mix_train,libri3mix_train,librispeech2mix_train,librispeech3mix_train,librispeech_train,libri2mix_train_gender,libri3mix_train_gender,librispeech2mix_train_gender,librispeech3mix_train_gender,libri2mix_train_keyword,libri3mix_train_keyword,librispeech2mix_train_keyword,librispeech3mix_train_keyword,librispeech2mix_train_order,librispeech3mix_train_order,de_train,de-en-2mix_train,de-en-3mix_train,de-en-2mix_train_targetLingual,de-en-3mix_train_targetLingual
100
+ valid_subset: librispeech2mix_test_targetASR,librispeech3mix_test_targetASR,librispeech2mix_test,librispeech3mix_test,librispeech_test_clean,librispeech2mix_test_gender,librispeech3mix_test_gender,librispeech2mix_test_keyword,librispeech3mix_test_keyword,librispeech2mix_test_order,librispeech3mix_test_order,de-en-2mix_test,de-en-3mix_test,de-en-2mix_test_targetLingual,de-en-3mix_test_targetLingual,de_test
101
+ combine_valid_subsets: null
102
+ ignore_unused_valid_subsets: false
103
+ validate_interval: 1
104
+ validate_interval_updates: 0
105
+ validate_after_updates: 0
106
+ fixed_validation_seed: null
107
+ disable_validation: true
108
+ max_tokens_valid: ${dataset.max_tokens}
109
+ batch_size_valid: ${dataset.batch_size}
110
+ max_valid_steps: null
111
+ curriculum: 0
112
+ gen_subset: test
113
+ num_shards: 1
114
+ shard_id: 0
115
+ grouped_shuffling: true
116
+ update_epoch_batch_itr: ${dataset.grouped_shuffling}
117
+ update_ordered_indices_seed: false
118
+ optimization:
119
+ _name: null
120
+ max_epoch: 0
121
+ max_update: 450000
122
+ stop_time_hours: 0.0
123
+ clip_norm: 1.0
124
+ sentence_avg: false
125
+ update_freq:
126
+ - 1
127
+ lr:
128
+ - 0.0001
129
+ stop_min_lr: -1.0
130
+ use_bmuf: false
131
+ skip_remainder_batch: false
132
+ checkpoint:
133
+ _name: null
134
+ save_dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
135
+ restore_file: checkpoint_last.pt
136
+ continue_once: null
137
+ finetune_from_model: null
138
+ reset_dataloader: false
139
+ reset_lr_scheduler: false
140
+ reset_meters: false
141
+ reset_optimizer: false
142
+ optimizer_overrides: '{}'
143
+ save_interval: 1
144
+ save_interval_updates: 10000
145
+ keep_interval_updates: 10
146
+ keep_interval_updates_pattern: -1
147
+ keep_last_epochs: -1
148
+ keep_best_checkpoints: -1
149
+ no_save: false
150
+ no_epoch_checkpoints: false
151
+ no_last_checkpoints: false
152
+ no_save_optimizer_state: false
153
+ best_checkpoint_metric: loss
154
+ maximize_best_checkpoint_metric: false
155
+ patience: -1
156
+ checkpoint_suffix: ''
157
+ checkpoint_shard_count: 1
158
+ load_checkpoint_on_all_dp_ranks: false
159
+ write_checkpoints_asynchronously: false
160
+ model_parallel_size: ${common.model_parallel_size}
161
+ bmuf:
162
+ _name: null
163
+ block_lr: 1.0
164
+ block_momentum: 0.875
165
+ global_sync_iter: 50
166
+ warmup_iterations: 500
167
+ use_nbm: false
168
+ average_sync: false
169
+ distributed_world_size: ${distributed_training.distributed_world_size}
170
+ generation:
171
+ _name: null
172
+ beam: 5
173
+ beam_mt: 0
174
+ nbest: 1
175
+ max_len_a: 0.0
176
+ max_len_b: 200
177
+ max_len_a_mt: 0.0
178
+ max_len_b_mt: 200
179
+ min_len: 1
180
+ match_source_len: false
181
+ unnormalized: false
182
+ no_early_stop: false
183
+ no_beamable_mm: false
184
+ lenpen: 1.0
185
+ lenpen_mt: 1.0
186
+ unkpen: 0.0
187
+ replace_unk: null
188
+ sacrebleu: false
189
+ score_reference: false
190
+ prefix_size: 0
191
+ no_repeat_ngram_size: 0
192
+ sampling: false
193
+ sampling_topk: -1
194
+ sampling_topp: -1.0
195
+ constraints: null
196
+ temperature: 1.0
197
+ diverse_beam_groups: -1
198
+ diverse_beam_strength: 0.5
199
+ diversity_rate: -1.0
200
+ print_alignment: null
201
+ print_step: false
202
+ lm_path: null
203
+ lm_weight: 0.0
204
+ iter_decode_eos_penalty: 0.0
205
+ iter_decode_max_iter: 10
206
+ iter_decode_force_max_iter: false
207
+ iter_decode_with_beam: 1
208
+ iter_decode_with_external_reranker: false
209
+ retain_iter_history: false
210
+ retain_dropout: false
211
+ retain_dropout_modules: null
212
+ decoding_format: null
213
+ no_seed_provided: false
214
+ eos_token: null
215
+ eval_lm:
216
+ _name: null
217
+ output_word_probs: false
218
+ output_word_stats: false
219
+ context_window: 0
220
+ softmax_batch: 9223372036854775807
221
+ interactive:
222
+ _name: null
223
+ buffer_size: 0
224
+ input: '-'
225
+ model:
226
+ _name: speechllm_model
227
+ freeze_audio_encoder: true
228
+ freeze_gpt: false
229
+ task:
230
+ _name: speechllm_task
231
+ data: /valleblob/v-lingmeng/speech/data/Whisper-Sidecar-data-metadata/data_for_wavllm
232
+ tokenizer_path: /valleblob/v-shujiehu/nlp/model/llama/tokenizer.model
233
+ processor_path: openai/whisper-large-v2
234
+ wavlm_processor_path: microsoft/wavlm-base
235
+ is_whisper: true
236
+ whisper_with_decoder: false
237
+ whisper_token_len: 64
238
+ freeze_audio_encoder: false
239
+ sft_stage: true
240
+ use_vicuna: false
241
+ llama_2: true
242
+ llama_2_path: /valleblob/v-shujiehu/nlp/model/llama-2-7b-chat/consolidated.00.pth
243
+ parallel_mode: false
244
+ use_lora: true
245
+ lora_r: 32
246
+ lora_alpha: 32
247
+ alpaca_text: true
248
+ prompt_bulid: true
249
+ use_wavlm: true
250
+ wavlm_weights: false
251
+ wavlm_output_weight: true
252
+ pretrained_checkpoint: /valleblob/v-shujiehu/checkpoints/speechllm/lora_32_32_wavlm_weighted_multitask_de_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_32gpu_1accum/checkpoint1.pt
253
+ criterion: cross_entropy_acc
254
+ optimizer:
255
+ _name: adam
256
+ adam_betas:
257
+ - 0.9
258
+ - 0.98
259
+ weight_decay: 0.01
260
+ lr_scheduler:
261
+ _name: inverse_sqrt
262
+ warmup_updates: 45000
263
+ warmup_init_lr: 1.0e-07
264
+ scoring: null
265
+ bpe: null
266
+ tokenizer: null
267
+ ema:
268
+ _name: null
269
+ store_ema: false
270
+ ema_decay: 0.9999
271
+ ema_start_update: 0
272
+ ema_seed_model: null
273
+ ema_update_freq: 1
274
+ ema_fp32: false
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/hydra.yaml ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
4
+ sweep:
5
+ dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: MULTIRUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.sweep.dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
114
+ - hydra.run.dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
115
+ - hydra.mode=MULTIRUN
116
+ task:
117
+ - common.user_dir=/home/aiscuser/code/speechllm/examples/speechllm/
118
+ - optimization.update_freq=[1]
119
+ - distributed_training.distributed_world_size=16
120
+ - common.tensorboard_logdir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
121
+ - checkpoint.save_dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
122
+ job:
123
+ name: hydra_train
124
+ chdir: null
125
+ override_dirname: checkpoint.save_dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum,common.tensorboard_logdir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum,common.user_dir=/home/aiscuser/code/speechllm/examples/speechllm/,distributed_training.distributed_world_size=16,optimization.update_freq=[1]
126
+ id: '0'
127
+ num: 0
128
+ config_name: speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.1'
139
+ cwd: /home/aiscuser/code/speechllm
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /home/aiscuser/code/speechllm/fairseq/config
145
+ schema: file
146
+ provider: main
147
+ - path: /home/aiscuser/code/speechllm/examples/speechllm/config/lingmeng
148
+ schema: file
149
+ provider: command-line
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0
154
+ choices:
155
+ hydra/env: default
156
+ hydra/callbacks: null
157
+ hydra/job_logging: default
158
+ hydra/hydra_logging: default
159
+ hydra/hydra_help: default
160
+ hydra/help: default
161
+ hydra/sweeper: basic
162
+ hydra/launcher: basic
163
+ hydra/output: default
164
+ verbose: false
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/.hydra/overrides.yaml ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ - common.user_dir=/home/aiscuser/code/speechllm/examples/speechllm/
2
+ - optimization.update_freq=[1]
3
+ - distributed_training.distributed_world_size=16
4
+ - common.tensorboard_logdir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
5
+ - checkpoint.save_dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/0/hydra_train.log ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-01-25 15:47:46,763][fairseq.distributed.utils][INFO] - distributed init (rank 0): tcp://node-0:9500
2
+ [2025-01-25 15:47:47,432][fairseq.distributed.utils][INFO] - distributed init (rank 2): tcp://node-0:9500
3
+ [2025-01-25 15:47:47,440][fairseq.distributed.utils][INFO] - distributed init (rank 5): tcp://node-0:9500
4
+ [2025-01-25 15:47:47,447][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 2
5
+ [2025-01-25 15:47:47,447][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 5
6
+ [2025-01-25 15:47:47,500][fairseq.distributed.utils][INFO] - distributed init (rank 7): tcp://node-0:9500
7
+ [2025-01-25 15:47:47,513][fairseq.distributed.utils][INFO] - distributed init (rank 3): tcp://node-0:9500
8
+ [2025-01-25 15:47:47,519][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 7
9
+ [2025-01-25 15:47:47,523][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 3
10
+ [2025-01-25 15:47:47,546][fairseq.distributed.utils][INFO] - distributed init (rank 4): tcp://node-0:9500
11
+ [2025-01-25 15:47:47,546][fairseq.distributed.utils][INFO] - distributed init (rank 1): tcp://node-0:9500
12
+ [2025-01-25 15:47:47,548][fairseq.distributed.utils][INFO] - distributed init (rank 6): tcp://node-0:9500
13
+ [2025-01-25 15:47:47,559][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 4
14
+ [2025-01-25 15:47:47,562][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 1
15
+ [2025-01-25 15:47:47,567][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 6
16
+ [2025-01-25 15:47:47,570][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 0
17
+ [2025-01-26 00:01:44,431][fairseq.distributed.utils][INFO] - distributed init (rank 8): tcp://node-0:9500
18
+ [2025-01-26 00:01:44,446][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 8
19
+ [2025-01-26 00:01:47,044][fairseq.distributed.utils][INFO] - distributed init (rank 14): tcp://node-0:9500
20
+ [2025-01-26 00:01:47,049][fairseq.distributed.utils][INFO] - distributed init (rank 15): tcp://node-0:9500
21
+ [2025-01-26 00:01:47,060][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 15
22
+ [2025-01-26 00:01:47,062][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 14
23
+ [2025-01-26 00:01:47,107][fairseq.distributed.utils][INFO] - distributed init (rank 13): tcp://node-0:9500
24
+ [2025-01-26 00:01:47,121][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 13
25
+ [2025-01-26 00:01:47,130][fairseq.distributed.utils][INFO] - distributed init (rank 9): tcp://node-0:9500
26
+ [2025-01-26 00:01:47,130][fairseq.distributed.utils][INFO] - distributed init (rank 10): tcp://node-0:9500
27
+ [2025-01-26 00:01:47,132][fairseq.distributed.utils][INFO] - distributed init (rank 11): tcp://node-0:9500
28
+ [2025-01-26 00:01:47,137][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 9
29
+ [2025-01-26 00:01:47,143][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 11
30
+ [2025-01-26 00:01:47,145][fairseq.distributed.utils][INFO] - distributed init (rank 12): tcp://node-0:9500
31
+ [2025-01-26 00:01:47,149][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 10
32
+ [2025-01-26 00:01:47,156][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 12
33
+ [2025-01-26 02:31:41,747][fairseq.distributed.utils][INFO] - distributed init (rank 0): tcp://node-0:9500
34
+ [2025-01-26 02:31:43,080][fairseq.distributed.utils][INFO] - distributed init (rank 7): tcp://node-0:9500
35
+ [2025-01-26 02:31:43,089][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 7
36
+ [2025-01-26 02:31:43,522][fairseq.distributed.utils][INFO] - distributed init (rank 3): tcp://node-0:9500
37
+ [2025-01-26 02:31:43,524][fairseq.distributed.utils][INFO] - distributed init (rank 5): tcp://node-0:9500
38
+ [2025-01-26 02:31:43,529][fairseq.distributed.utils][INFO] - distributed init (rank 4): tcp://node-0:9500
39
+ [2025-01-26 02:31:43,535][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 3
40
+ [2025-01-26 02:31:43,538][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 5
41
+ [2025-01-26 02:31:43,545][fairseq.distributed.utils][INFO] - distributed init (rank 6): tcp://node-0:9500
42
+ [2025-01-26 02:31:43,546][fairseq.distributed.utils][INFO] - distributed init (rank 2): tcp://node-0:9500
43
+ [2025-01-26 02:31:43,547][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 4
44
+ [2025-01-26 02:31:43,556][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 2
45
+ [2025-01-26 02:31:43,560][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 6
46
+ [2025-01-26 03:45:51,663][fairseq.distributed.utils][INFO] - distributed init (rank 8): tcp://node-0:9500
47
+ [2025-01-26 03:45:53,678][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 8
48
+ [2025-01-26 03:45:54,749][fairseq.distributed.utils][INFO] - distributed init (rank 12): tcp://node-0:9500
49
+ [2025-01-26 03:45:54,782][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 12
50
+ [2025-01-26 03:45:54,876][fairseq.distributed.utils][INFO] - distributed init (rank 14): tcp://node-0:9500
51
+ [2025-01-26 03:45:54,877][fairseq.distributed.utils][INFO] - distributed init (rank 13): tcp://node-0:9500
52
+ [2025-01-26 03:45:54,878][fairseq.distributed.utils][INFO] - distributed init (rank 11): tcp://node-0:9500
53
+ [2025-01-26 03:45:54,885][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 13
54
+ [2025-01-26 03:45:54,896][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 11
55
+ [2025-01-26 03:45:54,897][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 14
56
+ [2025-01-26 03:45:54,900][fairseq.distributed.utils][INFO] - distributed init (rank 10): tcp://node-0:9500
57
+ [2025-01-26 03:45:54,911][fairseq.distributed.utils][INFO] - distributed init (rank 15): tcp://node-0:9500
58
+ [2025-01-26 03:45:54,911][fairseq.distributed.utils][INFO] - distributed init (rank 9): tcp://node-0:9500
59
+ [2025-01-26 03:45:54,920][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 9
60
+ [2025-01-26 03:45:54,924][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 10
61
+ [2025-01-26 03:45:54,933][fairseq.distributed.utils][INFO] - initialized host node-1 as rank 15
62
+ [2025-01-27 23:14:33,771][fairseq.distributed.utils][INFO] - distributed init (rank 0): tcp://node-0:9500
63
+ [2025-01-27 23:14:33,996][fairseq.distributed.utils][INFO] - distributed init (rank 2): tcp://node-0:9500
64
+ [2025-01-27 23:14:34,012][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 2
65
+ [2025-01-27 23:14:34,063][fairseq.distributed.utils][INFO] - distributed init (rank 3): tcp://node-0:9500
66
+ [2025-01-27 23:14:34,080][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 3
67
+ [2025-01-27 23:14:34,157][fairseq.distributed.utils][INFO] - distributed init (rank 4): tcp://node-0:9500
68
+ [2025-01-27 23:14:34,164][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 4
69
+ [2025-01-27 23:14:34,303][fairseq.distributed.utils][INFO] - distributed init (rank 6): tcp://node-0:9500
70
+ [2025-01-27 23:14:34,320][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 6
71
+ [2025-01-27 23:14:34,329][fairseq.distributed.utils][INFO] - distributed init (rank 1): tcp://node-0:9500
72
+ [2025-01-27 23:14:34,330][fairseq.distributed.utils][INFO] - distributed init (rank 5): tcp://node-0:9500
73
+ [2025-01-27 23:14:34,331][fairseq.distributed.utils][INFO] - distributed init (rank 7): tcp://node-0:9500
74
+ [2025-01-27 23:14:34,338][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 7
75
+ [2025-01-27 23:14:34,340][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 1
76
+ [2025-01-27 23:14:34,345][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 5
77
+ [2025-01-27 23:14:34,519][fairseq.distributed.utils][INFO] - initialized host node-0 as rank 0
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/checkpoint1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:767be22c54b37f97539715f4e3abafb8ce99b38900ef5831d74749fa1d31bb04
3
+ size 30798558084
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/checkpoint2.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3cd0c0d579c5cd5efb0a82e304d3345e91c715ad92109551443788fb585c231
3
+ size 30798558084
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/multirun.yaml ADDED
@@ -0,0 +1,438 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
4
+ sweep:
5
+ dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: MULTIRUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.sweep.dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
114
+ - hydra.run.dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
115
+ - hydra.mode=MULTIRUN
116
+ task:
117
+ - common.user_dir=/home/aiscuser/code/speechllm/examples/speechllm/
118
+ - optimization.update_freq=[1]
119
+ - distributed_training.distributed_world_size=16
120
+ - common.tensorboard_logdir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
121
+ - checkpoint.save_dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
122
+ job:
123
+ name: hydra_train
124
+ chdir: null
125
+ override_dirname: checkpoint.save_dir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum,common.tensorboard_logdir=/valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum,common.user_dir=/home/aiscuser/code/speechllm/examples/speechllm/,distributed_training.distributed_world_size=16,optimization.update_freq=[1]
126
+ id: ???
127
+ num: ???
128
+ config_name: speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml
129
+ env_set: {}
130
+ env_copy: []
131
+ config:
132
+ override_dirname:
133
+ kv_sep: '='
134
+ item_sep: ','
135
+ exclude_keys: []
136
+ runtime:
137
+ version: 1.3.2
138
+ version_base: '1.1'
139
+ cwd: /home/aiscuser/code/speechllm
140
+ config_sources:
141
+ - path: hydra.conf
142
+ schema: pkg
143
+ provider: hydra
144
+ - path: /home/aiscuser/code/speechllm/fairseq/config
145
+ schema: file
146
+ provider: main
147
+ - path: /home/aiscuser/code/speechllm/examples/speechllm/config/lingmeng
148
+ schema: file
149
+ provider: command-line
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: ???
154
+ choices:
155
+ hydra/env: default
156
+ hydra/callbacks: null
157
+ hydra/job_logging: default
158
+ hydra/hydra_logging: default
159
+ hydra/hydra_help: default
160
+ hydra/help: default
161
+ hydra/sweeper: basic
162
+ hydra/launcher: basic
163
+ hydra/output: default
164
+ verbose: false
165
+ _name: null
166
+ common:
167
+ _name: null
168
+ no_progress_bar: false
169
+ log_interval: 100
170
+ log_format: json
171
+ log_file: null
172
+ aim_repo: null
173
+ aim_run_hash: null
174
+ tensorboard_logdir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
175
+ wandb_project: null
176
+ azureml_logging: false
177
+ seed: 100501
178
+ cpu: false
179
+ tpu: false
180
+ bf16: false
181
+ memory_efficient_bf16: false
182
+ fp16: true
183
+ memory_efficient_fp16: false
184
+ fp16_no_flatten_grads: false
185
+ fp16_init_scale: 128
186
+ fp16_scale_window: null
187
+ fp16_scale_tolerance: 0.02
188
+ on_cpu_convert_precision: false
189
+ min_loss_scale: 0.0001
190
+ threshold_loss_scale: null
191
+ amp: false
192
+ amp_batch_retries: 2
193
+ amp_init_scale: 128
194
+ amp_scale_window: null
195
+ user_dir: /home/aiscuser/code/speechllm/examples/speechllm/
196
+ empty_cache_freq: 0
197
+ all_gather_list_size: 16384
198
+ model_parallel_size: 1
199
+ quantization_config_path: null
200
+ profile: false
201
+ reset_logging: false
202
+ suppress_crashes: false
203
+ use_plasma_view: false
204
+ plasma_path: /tmp/plasma
205
+ common_eval:
206
+ _name: null
207
+ path: null
208
+ post_process: null
209
+ quiet: false
210
+ model_overrides: '{}'
211
+ results_path: null
212
+ distributed_training:
213
+ _name: null
214
+ distributed_world_size: 16
215
+ distributed_num_procs: 8
216
+ distributed_rank: 0
217
+ distributed_backend: nccl
218
+ distributed_init_method: null
219
+ distributed_port: -1
220
+ device_id: 7
221
+ distributed_no_spawn: false
222
+ ddp_backend: pytorch_ddp
223
+ ddp_comm_hook: none
224
+ bucket_cap_mb: 25
225
+ fix_batches_to_gpus: false
226
+ find_unused_parameters: true
227
+ gradient_as_bucket_view: false
228
+ fast_stat_sync: false
229
+ heartbeat_timeout: -1
230
+ broadcast_buffers: false
231
+ slowmo_momentum: null
232
+ slowmo_base_algorithm: localsgd
233
+ localsgd_frequency: 3
234
+ nprocs_per_node: 8
235
+ pipeline_model_parallel: false
236
+ pipeline_balance: null
237
+ pipeline_devices: null
238
+ pipeline_chunks: 0
239
+ pipeline_encoder_balance: null
240
+ pipeline_encoder_devices: null
241
+ pipeline_decoder_balance: null
242
+ pipeline_decoder_devices: null
243
+ pipeline_checkpoint: never
244
+ zero_sharding: none
245
+ fp16: ${common.fp16}
246
+ memory_efficient_fp16: ${common.memory_efficient_fp16}
247
+ tpu: ${common.tpu}
248
+ no_reshard_after_forward: false
249
+ fp32_reduce_scatter: false
250
+ cpu_offload: false
251
+ use_sharded_state: false
252
+ not_fsdp_flatten_parameters: false
253
+ dataset:
254
+ _name: null
255
+ num_workers: 0
256
+ skip_invalid_size_inputs_valid_test: true
257
+ max_tokens: 960000
258
+ batch_size: null
259
+ required_batch_size_multiple: 1
260
+ required_seq_len_multiple: 1
261
+ dataset_impl: null
262
+ data_buffer_size: 10
263
+ train_subset: libri2mix_train_targetASR,libri3mix_train_targetASR,librispeech2mix_train_targetASR,librispeech3mix_train_targetASR,libri2mix_train,libri3mix_train,librispeech2mix_train,librispeech3mix_train,librispeech_train,libri2mix_train_gender,libri3mix_train_gender,librispeech2mix_train_gender,librispeech3mix_train_gender,libri2mix_train_keyword,libri3mix_train_keyword,librispeech2mix_train_keyword,librispeech3mix_train_keyword,librispeech2mix_train_order,librispeech3mix_train_order,de_train,de-en-2mix_train,de-en-3mix_train,de-en-2mix_train_targetLingual,de-en-3mix_train_targetLingual
264
+ valid_subset: librispeech2mix_test_targetASR,librispeech3mix_test_targetASR,librispeech2mix_test,librispeech3mix_test,librispeech_test_clean,librispeech2mix_test_gender,librispeech3mix_test_gender,librispeech2mix_test_keyword,librispeech3mix_test_keyword,librispeech2mix_test_order,librispeech3mix_test_order,de-en-2mix_test,de-en-3mix_test,de-en-2mix_test_targetLingual,de-en-3mix_test_targetLingual,de_test
265
+ combine_valid_subsets: null
266
+ ignore_unused_valid_subsets: false
267
+ validate_interval: 1
268
+ validate_interval_updates: 0
269
+ validate_after_updates: 0
270
+ fixed_validation_seed: null
271
+ disable_validation: true
272
+ max_tokens_valid: ${dataset.max_tokens}
273
+ batch_size_valid: ${dataset.batch_size}
274
+ max_valid_steps: null
275
+ curriculum: 0
276
+ gen_subset: test
277
+ num_shards: 1
278
+ shard_id: 0
279
+ grouped_shuffling: true
280
+ update_epoch_batch_itr: ${dataset.grouped_shuffling}
281
+ update_ordered_indices_seed: false
282
+ optimization:
283
+ _name: null
284
+ max_epoch: 0
285
+ max_update: 450000
286
+ stop_time_hours: 0.0
287
+ clip_norm: 1.0
288
+ sentence_avg: false
289
+ update_freq:
290
+ - 1
291
+ lr:
292
+ - 0.0001
293
+ stop_min_lr: -1.0
294
+ use_bmuf: false
295
+ skip_remainder_batch: false
296
+ checkpoint:
297
+ _name: null
298
+ save_dir: /valleblob/v-lingmeng/speech/model/speechllm_multispk/lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum
299
+ restore_file: checkpoint_last.pt
300
+ continue_once: null
301
+ finetune_from_model: null
302
+ reset_dataloader: false
303
+ reset_lr_scheduler: false
304
+ reset_meters: false
305
+ reset_optimizer: false
306
+ optimizer_overrides: '{}'
307
+ save_interval: 1
308
+ save_interval_updates: 10000
309
+ keep_interval_updates: 10
310
+ keep_interval_updates_pattern: -1
311
+ keep_last_epochs: -1
312
+ keep_best_checkpoints: -1
313
+ no_save: false
314
+ no_epoch_checkpoints: false
315
+ no_last_checkpoints: false
316
+ no_save_optimizer_state: false
317
+ best_checkpoint_metric: loss
318
+ maximize_best_checkpoint_metric: false
319
+ patience: -1
320
+ checkpoint_suffix: ''
321
+ checkpoint_shard_count: 1
322
+ load_checkpoint_on_all_dp_ranks: false
323
+ write_checkpoints_asynchronously: false
324
+ model_parallel_size: ${common.model_parallel_size}
325
+ bmuf:
326
+ _name: null
327
+ block_lr: 1.0
328
+ block_momentum: 0.875
329
+ global_sync_iter: 50
330
+ warmup_iterations: 500
331
+ use_nbm: false
332
+ average_sync: false
333
+ distributed_world_size: ${distributed_training.distributed_world_size}
334
+ generation:
335
+ _name: null
336
+ beam: 5
337
+ beam_mt: 0
338
+ nbest: 1
339
+ max_len_a: 0.0
340
+ max_len_b: 200
341
+ max_len_a_mt: 0.0
342
+ max_len_b_mt: 200
343
+ min_len: 1
344
+ match_source_len: false
345
+ unnormalized: false
346
+ no_early_stop: false
347
+ no_beamable_mm: false
348
+ lenpen: 1.0
349
+ lenpen_mt: 1.0
350
+ unkpen: 0.0
351
+ replace_unk: null
352
+ sacrebleu: false
353
+ score_reference: false
354
+ prefix_size: 0
355
+ no_repeat_ngram_size: 0
356
+ sampling: false
357
+ sampling_topk: -1
358
+ sampling_topp: -1.0
359
+ constraints: null
360
+ temperature: 1.0
361
+ diverse_beam_groups: -1
362
+ diverse_beam_strength: 0.5
363
+ diversity_rate: -1.0
364
+ print_alignment: null
365
+ print_step: false
366
+ lm_path: null
367
+ lm_weight: 0.0
368
+ iter_decode_eos_penalty: 0.0
369
+ iter_decode_max_iter: 10
370
+ iter_decode_force_max_iter: false
371
+ iter_decode_with_beam: 1
372
+ iter_decode_with_external_reranker: false
373
+ retain_iter_history: false
374
+ retain_dropout: false
375
+ retain_dropout_modules: null
376
+ decoding_format: null
377
+ no_seed_provided: false
378
+ eos_token: null
379
+ eval_lm:
380
+ _name: null
381
+ output_word_probs: false
382
+ output_word_stats: false
383
+ context_window: 0
384
+ softmax_batch: 9223372036854775807
385
+ interactive:
386
+ _name: null
387
+ buffer_size: 0
388
+ input: '-'
389
+ model:
390
+ _name: speechllm_model
391
+ freeze_audio_encoder: true
392
+ freeze_gpt: false
393
+ task:
394
+ _name: speechllm_task
395
+ data: /valleblob/v-lingmeng/speech/data/Whisper-Sidecar-data-metadata/data_for_wavllm
396
+ tokenizer_path: /valleblob/v-shujiehu/nlp/model/llama/tokenizer.model
397
+ processor_path: openai/whisper-large-v2
398
+ wavlm_processor_path: microsoft/wavlm-base
399
+ is_whisper: true
400
+ whisper_with_decoder: false
401
+ whisper_token_len: 64
402
+ freeze_audio_encoder: false
403
+ sft_stage: true
404
+ use_vicuna: false
405
+ llama_2: true
406
+ llama_2_path: /valleblob/v-shujiehu/nlp/model/llama-2-7b-chat/consolidated.00.pth
407
+ parallel_mode: false
408
+ use_lora: true
409
+ lora_r: 32
410
+ lora_alpha: 32
411
+ alpaca_text: true
412
+ prompt_bulid: true
413
+ use_wavlm: true
414
+ wavlm_weights: false
415
+ wavlm_output_weight: true
416
+ pretrained_checkpoint: /valleblob/v-shujiehu/checkpoints/speechllm/lora_32_32_wavlm_weighted_multitask_de_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_32gpu_1accum/checkpoint1.pt
417
+ criterion: cross_entropy_acc
418
+ optimizer:
419
+ _name: adam
420
+ adam_betas:
421
+ - 0.9
422
+ - 0.98
423
+ weight_decay: 0.01
424
+ lr_scheduler:
425
+ _name: inverse_sqrt
426
+ warmup_updates: 45000
427
+ warmup_init_lr: 1.0e-07
428
+ scoring: null
429
+ bpe: null
430
+ tokenizer: null
431
+ ema:
432
+ _name: null
433
+ store_ema: false
434
+ ema_decay: 0.9999
435
+ ema_start_update: 0
436
+ ema_seed_model: null
437
+ ema_update_freq: 1
438
+ ema_fp32: false
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train.log ADDED
Binary file (13.2 kB). View file
 
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train/events.out.tfevents.1738168057.node-0.6519.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56020f06275ed4c2b627abccd75c1d7e09cb77e6bd9f94f17ac4223f79465a29
3
+ size 2907
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train_inner/events.out.tfevents.1737792645.node-0.6570.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de30971eca79360a9ffe202c2145dd63fd81ef7d3193ce8ecc0dd0495407ed7b
3
+ size 42074
lingmeng_multispk_multitask_retrain_speechllm_v0.1_llama2_chat_wavlm_weighted_update_lora_32_32_prompt_build_multispk_multitask_de.yaml_16gpu_1accum/train_inner/events.out.tfevents.1737992273.node-0.6519.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37d6fb6f81f6107030e3d92babcd5b96d9034cc1ac1fcb169f2bc84ba190516d
3
+ size 2072101