Jakob Poncelet commited on
Commit
150f2e5
·
1 Parent(s): e1de3b7

alternative model version

Browse files
.gitattributes CHANGED
@@ -35,3 +35,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  model.pth filter=lfs diff=lfs merge=lfs -text
37
  bpe.model filter=lfs diff=lfs merge=lfs -text
 
 
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  model.pth filter=lfs diff=lfs merge=lfs -text
37
  bpe.model filter=lfs diff=lfs merge=lfs -text
38
+ old_model/model.pth filter=lfs diff=lfs merge=lfs -text
model.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8e3a313e405292897dbe70875d084c7facce1f36e924e7b6793b930bb3a3010e
3
- size 723062244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fd2f27c419742d03f205dc5d38364ab5c9c892f4b0bfdfe510244ce92073dea
3
+ size 723052874
old_model/model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e3a313e405292897dbe70875d084c7facce1f36e924e7b6793b930bb3a3010e
3
+ size 723062244
old_model/train.yaml ADDED
The diff for this file is too large to render. See raw diff
 
old_model/train_config.yaml ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ batch_size: 16
2
+ accum_grad: 128
3
+ max_epoch: 150
4
+ patience: none
5
+ # The initialization method for model parameters
6
+ init: xavier_uniform
7
+ best_model_criterion:
8
+ - - valid
9
+ - acc_asr #
10
+ - max
11
+ keep_nbest_models: 10
12
+
13
+ encoder: conformer
14
+ encoder_conf:
15
+ output_size: 512
16
+ attention_heads: 8
17
+ linear_units: 2048
18
+ num_blocks: 12
19
+ dropout_rate: 0.1
20
+ positional_dropout_rate: 0.1
21
+ attention_dropout_rate: 0.1
22
+ input_layer: conv2d
23
+ normalize_before: true
24
+ macaron_style: true
25
+ rel_pos_type: latest
26
+ pos_enc_layer_type: rel_pos
27
+ selfattention_layer_type: rel_selfattn
28
+ activation_type: swish
29
+ use_cnn_module: true
30
+ cnn_module_kernel: 31
31
+ interctc_layer_idx: [6]
32
+
33
+ decoder: multi_transformer #mlm
34
+ decoder_conf:
35
+ attention_heads: 8
36
+ linear_units: 2048
37
+ num_blocks: 6
38
+ dropout_rate: 0.1
39
+ positional_dropout_rate: 0.1
40
+ self_attention_dropout_rate: 0.1
41
+ src_attention_dropout_rate: 0.1
42
+ return_hidden: true
43
+
44
+ subtitle_encoder: transformer
45
+ subtitle_encoder_conf:
46
+ output_size: 512
47
+ attention_heads: 8
48
+ linear_units: 2048
49
+ num_blocks: 6
50
+ dropout_rate: 0.1
51
+ positional_dropout_rate: 0.1
52
+ attention_dropout_rate: 0.1
53
+ input_layer: none
54
+ normalize_before: true
55
+ macaron_style: true
56
+
57
+ subtitle_decoder: multi_transformer
58
+ subtitle_decoder_conf:
59
+ attention_heads: 8
60
+ linear_units: 2048
61
+ num_blocks: 6
62
+ dropout_rate: 0.1
63
+ positional_dropout_rate: 0.1
64
+ self_attention_dropout_rate: 0.1
65
+ src_attention_dropout_rate: 0.1
66
+
67
+ model_conf:
68
+ asr_weight: 0.5
69
+ subs_weight: 0.5
70
+ ctc_weight: 0.3 # = CTC weight
71
+ interctc_weight: 0.3
72
+ lsm_weight_asr: 0.1
73
+ lsm_weight_mt: 0.1
74
+ length_normalized_loss: false
75
+ condition_subtitle_decoder: true
76
+ condition_asr_decoder: true
77
+ use_asr_feats: "encoder"
78
+
79
+ optim: adam
80
+ optim_conf:
81
+ lr: 0.003
82
+ scheduler: warmuplr
83
+ scheduler_conf:
84
+ warmup_steps: 100000
85
+
86
+ specaug: specaug
87
+ specaug_conf:
88
+ apply_time_warp: true
89
+ time_warp_window: 5
90
+ time_warp_mode: bicubic
91
+ apply_freq_mask: true
92
+ freq_mask_width_range:
93
+ - 0
94
+ - 30
95
+ num_freq_mask: 2
96
+ apply_time_mask: true
97
+ time_mask_width_range:
98
+ - 0
99
+ - 40
100
+ num_time_mask: 2
101
+
train.yaml CHANGED
@@ -4,13 +4,13 @@ log_level: INFO
4
  dry_run: false
5
  iterator_type: sequence
6
  batch_asr_ratio: null
7
- output_dir: /esat/spchtemp/scratch/jponcele/espnet2/exp/exp-hpc/st_train_subtitling_chained_PL_C10_new_combined_nelf_all_dim512_iter10k
8
  ngpu: 1
9
  seed: 0
10
  num_workers: 1
11
  num_att_plot: 0
12
  dist_backend: nccl
13
- dist_init_method: file:///dodrio/scratch/projects/starting_2023_035/experiments/espnet2/exp/exp-new-combined/st_train_subtitling_chained_PL_C8_new_combined_nelf_all_dim512_iter10k/.dist_init_eba45741-b85b-43d0-b471-bda370ed01d6
14
  dist_world_size: 16
15
  dist_rank: 0
16
  local_rank: 0
@@ -10145,6 +10145,7 @@ model_conf:
10145
  condition_subtitle_decoder: true
10146
  condition_asr_decoder: true
10147
  use_asr_feats: encoder
 
10148
  use_preprocessor: true
10149
  token_type: bpe
10150
  src_token_type: bpe
 
4
  dry_run: false
5
  iterator_type: sequence
6
  batch_asr_ratio: null
7
+ output_dir: /esat/spchtemp/scratch/jponcele/espnet2/exp/exp-hpc/st_train_subtitling_chained_PL_C10_new_combined_nelf_all_dim512_fix_conditioning_iter10k
8
  ngpu: 1
9
  seed: 0
10
  num_workers: 1
11
  num_att_plot: 0
12
  dist_backend: nccl
13
+ dist_init_method: file:///dodrio/scratch/projects/starting_2023_035/experiments/espnet2/exp/exp-new-combined/st_train_subtitling_chained_PL_C8_new_combined_nelf_all_dim512_fix_conditioning_iter10k/.dist_init_8e4c5c43-7800-495a-a33a-312f06960745
14
  dist_world_size: 16
15
  dist_rank: 0
16
  local_rank: 0
 
10145
  condition_subtitle_decoder: true
10146
  condition_asr_decoder: true
10147
  use_asr_feats: encoder
10148
+ fix_conditioning: true
10149
  use_preprocessor: true
10150
  token_type: bpe
10151
  src_token_type: bpe
train_config.yaml CHANGED
@@ -74,6 +74,7 @@ model_conf:
74
  condition_subtitle_decoder: true
75
  condition_asr_decoder: true
76
  use_asr_feats: "encoder"
 
77
 
78
  optim: adam
79
  optim_conf:
 
74
  condition_subtitle_decoder: true
75
  condition_asr_decoder: true
76
  use_asr_feats: "encoder"
77
+ fix_conditioning: true
78
 
79
  optim: adam
80
  optim_conf: