mrq
commited on
Commit
·
4edce83
1
Parent(s):
532d199
YAML
Browse files- config.yaml +9 -31
config.yaml
CHANGED
|
@@ -20,41 +20,22 @@ hyperparameters:
|
|
| 20 |
batch_size: 32
|
| 21 |
gradient_accumulation_steps: 4
|
| 22 |
gradient_clipping: 100
|
| 23 |
-
|
| 24 |
optimizer: AdamW
|
| 25 |
learning_rate: 1.0e-6
|
| 26 |
-
|
| 27 |
-
scheduler_type: ""
|
| 28 |
-
#scheduler_type: OneCycle
|
| 29 |
-
#scheduler_params:
|
| 30 |
-
# cycle_first_step_size: 10_000
|
| 31 |
-
# cycle_first_stair_count: 10_000
|
| 32 |
-
|
| 33 |
-
# cycle_second_step_size: 15_000
|
| 34 |
-
# cycle_second_stair_count: 15_000
|
| 35 |
-
|
| 36 |
-
# decay_step_size: 5_000
|
| 37 |
-
|
| 38 |
-
# cycle_min_lr: 2.5e-4 # 1.0e-5
|
| 39 |
-
# cycle_max_lr: 2.5e-4 # 1.0e-4
|
| 40 |
-
# decay_lr_rate: 0.0
|
| 41 |
-
|
| 42 |
-
# cycle_min_mom: 0.90
|
| 43 |
-
# cycle_max_mom: 0.99
|
| 44 |
-
# decay_mom_rate: 0.0
|
| 45 |
|
| 46 |
evaluation:
|
| 47 |
batch_size: 64
|
| 48 |
frequency: 500
|
| 49 |
size: 64
|
| 50 |
-
|
| 51 |
steps: 300
|
| 52 |
ar_temperature: 0.95
|
| 53 |
nar_temperature: 0.25
|
| 54 |
|
| 55 |
trainer:
|
| 56 |
iterations: 1_000_000
|
| 57 |
-
|
| 58 |
save_tag: step
|
| 59 |
save_on_oom: True
|
| 60 |
save_on_quit: True
|
|
@@ -66,17 +47,14 @@ trainer:
|
|
| 66 |
|
| 67 |
load_state_dict: True
|
| 68 |
strict_loading: False
|
| 69 |
-
|
| 70 |
-
#load_states: False
|
| 71 |
-
#restart_step_count: True
|
| 72 |
-
|
| 73 |
gc_mode: None # "global_step"
|
| 74 |
|
| 75 |
-
weight_dtype:
|
| 76 |
|
| 77 |
-
backend:
|
| 78 |
deepspeed:
|
| 79 |
-
zero_optimization_level:
|
| 80 |
use_compression_training: True
|
| 81 |
|
| 82 |
inference:
|
|
@@ -93,7 +71,7 @@ bitsandbytes:
|
|
| 93 |
|
| 94 |
dataset:
|
| 95 |
speaker_name_getter: "lambda p: f'{p.parts[-3]}_{p.parts[-2]}'"
|
| 96 |
-
|
| 97 |
use_hdf5: True
|
| 98 |
hdf5_flag: r
|
| 99 |
validate: True
|
|
@@ -3229,4 +3207,4 @@ dataset:
|
|
| 3229 |
]
|
| 3230 |
noise: [
|
| 3231 |
"/data/Other/noise",
|
| 3232 |
-
]
|
|
|
|
| 20 |
batch_size: 32
|
| 21 |
gradient_accumulation_steps: 4
|
| 22 |
gradient_clipping: 100
|
| 23 |
+
|
| 24 |
optimizer: AdamW
|
| 25 |
learning_rate: 1.0e-6
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
evaluation:
|
| 28 |
batch_size: 64
|
| 29 |
frequency: 500
|
| 30 |
size: 64
|
| 31 |
+
|
| 32 |
steps: 300
|
| 33 |
ar_temperature: 0.95
|
| 34 |
nar_temperature: 0.25
|
| 35 |
|
| 36 |
trainer:
|
| 37 |
iterations: 1_000_000
|
| 38 |
+
|
| 39 |
save_tag: step
|
| 40 |
save_on_oom: True
|
| 41 |
save_on_quit: True
|
|
|
|
| 47 |
|
| 48 |
load_state_dict: True
|
| 49 |
strict_loading: False
|
| 50 |
+
|
|
|
|
|
|
|
|
|
|
| 51 |
gc_mode: None # "global_step"
|
| 52 |
|
| 53 |
+
weight_dtype: float32
|
| 54 |
|
| 55 |
+
backend: local
|
| 56 |
deepspeed:
|
| 57 |
+
zero_optimization_level: 0
|
| 58 |
use_compression_training: True
|
| 59 |
|
| 60 |
inference:
|
|
|
|
| 71 |
|
| 72 |
dataset:
|
| 73 |
speaker_name_getter: "lambda p: f'{p.parts[-3]}_{p.parts[-2]}'"
|
| 74 |
+
|
| 75 |
use_hdf5: True
|
| 76 |
hdf5_flag: r
|
| 77 |
validate: True
|
|
|
|
| 3207 |
]
|
| 3208 |
noise: [
|
| 3209 |
"/data/Other/noise",
|
| 3210 |
+
]
|