Respair commited on
Commit
08cd977
·
verified ·
1 Parent(s): 7e5c8d4

Create fast-conformer_aed.yaml

Browse files
Files changed (1) hide show
  1. config/fast-conformer_aed.yaml +51 -15
config/fast-conformer_aed.yaml CHANGED
@@ -22,14 +22,25 @@ name: "FastConformer-Transformer-MultiTask"
22
 
23
  # Note: for larger models (1B+ params) initializing from a pretrained encoder
24
  # may help (or even be required to) stabilize the training.
25
- init_from_nemo_model: null
 
 
 
 
 
 
 
 
 
 
 
26
 
27
  # If using example training script, below will be used to instantiate spl_tokens tokenizer.
28
  # Similar can be done by calling CanaryTokenizer.build_special_tokenizer(tokens, output_dir).
29
  # If a tokenizer exists in dir, will skip building and use already built tokenizer.
30
  spl_tokens:
31
  model_dir: ???
32
- tokens: ["translate", "transcribe", "ja"]
33
  force_rebuild: False # Set to True to build new tokenizer each time.
34
 
35
  model:
@@ -229,7 +240,7 @@ model:
229
  return_best_hypothesis: true # Returns the most probably hypothesis after beam search
230
 
231
  beam:
232
- beam_size: 1
233
  len_pen: 0.0
234
  max_generation_delta: 50
235
 
@@ -253,7 +264,7 @@ model:
253
  sched:
254
  name: InverseSquareRootAnnealing
255
  # scheduler config override
256
- warmup_steps: 2500
257
  warmup_ratio: null
258
  min_lr: 1e-6
259
 
@@ -279,27 +290,52 @@ trainer:
279
  logger: false # Provided by exp_manager
280
  use_distributed_sampler: false # Lhotse has its own distributed sampler
281
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
282
  exp_manager:
283
  exp_dir: null
284
  name: ${name}
285
  create_tensorboard_logger: true
 
286
  create_checkpoint_callback: true
287
  checkpoint_callback_params:
288
- # in case of multiple validation sets, first one is used
289
- monitor: "val_loss"
 
290
  mode: "min"
291
- save_top_k: 5
292
- always_save_nemo: True # saves the checkpoints as nemo files instead of PTL checkpoints
 
 
 
293
  # checkpoint_callback_params:
294
- # every_n_train_steps: 2000
295
- # every_n_epochs: null # must be set to null to use every_n_train_steps
296
- # monitor: "step" # want all checkpoints, so step + mode: max always succeeds
297
  # mode: "min"
298
- # save_top_k: 5 # save all models
299
- # save_last: True
300
- # always_save_nemo: True
 
301
 
302
- resume_from_checkpoint: "/home/ubuntu/NeMo/canary_results/canary-small/checkpoints/canary-small--val_loss=0.1680-epoch=16.ckpt" # The path to a checkpoint file to continue the training, restores the whole state including the epoch, step, LR schedulers, apex, etc.
303
  # you need to set these two to True to continue the training
304
  resume_if_exists: true
305
  resume_ignore_no_checkpoint: true
 
22
 
23
  # Note: for larger models (1B+ params) initializing from a pretrained encoder
24
  # may help (or even be required to) stabilize the training.
25
+
26
+ init_from_nemo_model:
27
+
28
+ model0:
29
+ path: "/home/ubuntu/NeMo_Canary/canary_results/Higurashi_ASR/checkpoints/Higurashi_ASR.nemo"
30
+ exclude: ["transf_decoder._embedding.token_embedding", "log_softmax.mlp.layer0"]
31
+
32
+ # init_from_pretrained_model:
33
+ # model0:
34
+ # name: "nvidia/canary-180m-flash"
35
+ # include: ["encoder"]
36
+
37
 
38
  # If using example training script, below will be used to instantiate spl_tokens tokenizer.
39
  # Similar can be done by calling CanaryTokenizer.build_special_tokenizer(tokens, output_dir).
40
  # If a tokenizer exists in dir, will skip building and use already built tokenizer.
41
  spl_tokens:
42
  model_dir: ???
43
+ tokens: ["translate", "transcribe", 'ja']
44
  force_rebuild: False # Set to True to build new tokenizer each time.
45
 
46
  model:
 
240
  return_best_hypothesis: true # Returns the most probably hypothesis after beam search
241
 
242
  beam:
243
+ beam_size: 4
244
  len_pen: 0.0
245
  max_generation_delta: 50
246
 
 
264
  sched:
265
  name: InverseSquareRootAnnealing
266
  # scheduler config override
267
+ warmup_steps: 5000
268
  warmup_ratio: null
269
  min_lr: 1e-6
270
 
 
290
  logger: false # Provided by exp_manager
291
  use_distributed_sampler: false # Lhotse has its own distributed sampler
292
 
293
+ # exp_manager:
294
+ # exp_dir: null
295
+ # name: ${name}
296
+ # create_tensorboard_logger: true
297
+ # create_checkpoint_callback: true
298
+ # checkpoint_callback_params:
299
+ # # in case of multiple validation sets, first one is used
300
+ # monitor: "val_loss"
301
+ # mode: "min"
302
+ # save_top_k: 5
303
+ # always_save_nemo: True # saves the checkpoints as nemo files instead of PTL checkpoints
304
+ # # checkpoint_callback_params:
305
+ # # every_n_train_steps: 2000
306
+ # # every_n_epochs: null # must be set to null to use every_n_train_steps
307
+ # # monitor: "val_loss" # want all checkpoints, so step + mode: max always succeeds
308
+ # # mode: "min"
309
+ # # save_top_k: 5 # save all models
310
+ # # save_last: True
311
+ # # always_save_nemo: True
312
+
313
  exp_manager:
314
  exp_dir: null
315
  name: ${name}
316
  create_tensorboard_logger: true
317
+
318
  create_checkpoint_callback: true
319
  checkpoint_callback_params:
320
+ every_n_train_steps: 4990
321
+ every_n_epochs: null # must be set to null to use every_n_train_steps
322
+ monitor: "step" # want all checkpoints, so step + mode: max always succeeds
323
  mode: "min"
324
+ save_top_k: 5 # save all models
325
+ save_last: True
326
+ always_save_nemo: True
327
+
328
+ # create_checkpoint_callback: true
329
  # checkpoint_callback_params:
330
+ # # in case of multiple validation sets, first one is used
331
+ # monitor: "val_loss"
 
332
  # mode: "min"
333
+ # save_top_k: 5
334
+ # always_save_nemo: True # saves the checkpoints as nemo files instead of PTL checkpoints
335
+
336
+
337
 
338
+ resume_from_checkpoint: /home/ubuntu/NeMo_Canary/canary_results/Higurashi_ASR_v.02/checkpoints/Higurashi_ASR_v.02--step=29940.0000-epoch=1-last.ckpt # The path to a checkpoint file to continue the training, restores the whole state including the epoch, step, LR schedulers, apex, etc.
339
  # you need to set these two to True to continue the training
340
  resume_if_exists: true
341
  resume_ignore_no_checkpoint: true