File size: 2,191 Bytes
297c44b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
# Generated 2021-10-05 from:
# /mnt/md0/user_wayne/speechbrain/recipes/MATBN/LM/hparams/TransformerLM_seg_char.yaml
# yamllint disable
output_folder: results/TransformerLM_seg_char
save_folder: results/TransformerLM_seg_char/save
train_log: results/TransformerLM_seg_char/train_log.txt
num_workers: 4

data_folder: results/prepare_seg

tokenizer_file: results/tokenizer_seg_bpe5k_char/5000_char.model

tokenizer: &id001 !new:sentencepiece.SentencePieceProcessor

pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
  collect_in: results/TransformerLM_seg_char/tokenizer
  loadables:
    tokenizer: *id001
  paths:
    tokenizer: results/tokenizer_seg_bpe5k_char/5000_char.model

train_logger: !new:speechbrain.utils.train_logger.FileTrainLogger
  save_file: results/TransformerLM_seg_char/train_log.txt

# Training parameters
number_of_epochs: 20
batch_size: 64
lr: 1
accumulation_steps: 2
ckpt_interval_minutes: 15

epoch_counter: &id004 !new:speechbrain.utils.epoch_loop.EpochCounter

  limit: 20

# Dataloader options
train_dataloader_opts:
  batch_size: 64
  num_workers: 4
  shuffle: true
  pin_memory: true

valid_dataloader_opts:
  batch_size: 64
  num_workers: 4

test_dataloader_opts:
  batch_size: 64
  num_workers: 4

# Model parameters
d_model: 576

# Outputs
output_neurons: 5000
blank_index: 0
bos_index: 1
eos_index: 2
unk_index: 0
pad_index: 0

model: &id002 !new:speechbrain.lobes.models.transformer.TransformerLM.TransformerLM

  vocab: 5000
  d_model: 576
  nhead: 6
  num_encoder_layers: 6
  num_decoder_layers: 0
  d_ffn: 1538
  dropout: 0.2
  activation: !name:torch.nn.GELU
  normalize_before: false

modules:
  model: *id002
lr_annealing: &id003 !new:speechbrain.nnet.schedulers.NoamScheduler
  lr_initial: 1
  n_warmup_steps: 1000
  model_size: 576

checkpointer: !new:speechbrain.utils.checkpoints.Checkpointer
  checkpoints_dir: results/TransformerLM_seg_char/save
  recoverables:
    model: *id002
    scheduler: *id003
    counter: *id004
log_softmax: !new:speechbrain.nnet.activations.Softmax
  apply_log: true

optimizer: !name:torch.optim.Adam
  lr: 0
  betas: (0.9, 0.98)
  eps: 0.000000001

compute_cost: !name:speechbrain.nnet.losses.nll_loss