Spaces:
Running
on
Zero
Running
on
Zero
Delete configs
Browse files- configs/s1.yaml +0 -31
- configs/s1big.yaml +0 -31
- configs/s1big2.yaml +0 -31
- configs/s1longer-v2.yaml +0 -31
- configs/s1longer.yaml +0 -31
- configs/s1mq.yaml +0 -77
- configs/s2.json +0 -90
- configs/train.yaml +0 -32
configs/s1.yaml
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 300
|
| 4 |
-
batch_size: 8
|
| 5 |
-
gradient_accumulation: 4
|
| 6 |
-
save_every_n_epoch: 1
|
| 7 |
-
precision: 16
|
| 8 |
-
gradient_clip: 1.0
|
| 9 |
-
optimizer:
|
| 10 |
-
lr: 0.01
|
| 11 |
-
lr_init: 0.00001
|
| 12 |
-
lr_end: 0.0001
|
| 13 |
-
warmup_steps: 2000
|
| 14 |
-
decay_steps: 40000
|
| 15 |
-
data:
|
| 16 |
-
max_eval_sample: 8
|
| 17 |
-
max_sec: 54
|
| 18 |
-
num_workers: 1
|
| 19 |
-
pad_val: 1024 # same with EOS in model
|
| 20 |
-
model:
|
| 21 |
-
vocab_size: 1025
|
| 22 |
-
phoneme_vocab_size: 512
|
| 23 |
-
embedding_dim: 512
|
| 24 |
-
hidden_dim: 512
|
| 25 |
-
head: 16
|
| 26 |
-
linear_units: 2048
|
| 27 |
-
n_layer: 12
|
| 28 |
-
dropout: 0
|
| 29 |
-
EOS: 1024
|
| 30 |
-
inference:
|
| 31 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s1big.yaml
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 300
|
| 4 |
-
batch_size: 8
|
| 5 |
-
gradient_accumulation: 4
|
| 6 |
-
save_every_n_epoch: 1
|
| 7 |
-
precision: 16-mixed
|
| 8 |
-
gradient_clip: 1.0
|
| 9 |
-
optimizer:
|
| 10 |
-
lr: 0.01
|
| 11 |
-
lr_init: 0.00001
|
| 12 |
-
lr_end: 0.0001
|
| 13 |
-
warmup_steps: 2000
|
| 14 |
-
decay_steps: 40000
|
| 15 |
-
data:
|
| 16 |
-
max_eval_sample: 8
|
| 17 |
-
max_sec: 54
|
| 18 |
-
num_workers: 1
|
| 19 |
-
pad_val: 1024 # same with EOS in model
|
| 20 |
-
model:
|
| 21 |
-
vocab_size: 1025
|
| 22 |
-
phoneme_vocab_size: 512
|
| 23 |
-
embedding_dim: 1024
|
| 24 |
-
hidden_dim: 1024
|
| 25 |
-
head: 16
|
| 26 |
-
linear_units: 2048
|
| 27 |
-
n_layer: 16
|
| 28 |
-
dropout: 0
|
| 29 |
-
EOS: 1024
|
| 30 |
-
inference:
|
| 31 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s1big2.yaml
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 300
|
| 4 |
-
batch_size: 12
|
| 5 |
-
gradient_accumulation: 4
|
| 6 |
-
save_every_n_epoch: 1
|
| 7 |
-
precision: 16-mixed
|
| 8 |
-
gradient_clip: 1.0
|
| 9 |
-
optimizer:
|
| 10 |
-
lr: 0.01
|
| 11 |
-
lr_init: 0.00001
|
| 12 |
-
lr_end: 0.0001
|
| 13 |
-
warmup_steps: 2000
|
| 14 |
-
decay_steps: 40000
|
| 15 |
-
data:
|
| 16 |
-
max_eval_sample: 8
|
| 17 |
-
max_sec: 54
|
| 18 |
-
num_workers: 1
|
| 19 |
-
pad_val: 1024 # same with EOS in model
|
| 20 |
-
model:
|
| 21 |
-
vocab_size: 1025
|
| 22 |
-
phoneme_vocab_size: 512
|
| 23 |
-
embedding_dim: 1024
|
| 24 |
-
hidden_dim: 1024
|
| 25 |
-
head: 16
|
| 26 |
-
linear_units: 2048
|
| 27 |
-
n_layer: 6
|
| 28 |
-
dropout: 0
|
| 29 |
-
EOS: 1024
|
| 30 |
-
inference:
|
| 31 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s1longer-v2.yaml
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 20
|
| 4 |
-
batch_size: 8
|
| 5 |
-
save_every_n_epoch: 1
|
| 6 |
-
precision: 16-mixed
|
| 7 |
-
gradient_clip: 1.0
|
| 8 |
-
optimizer:
|
| 9 |
-
lr: 0.01
|
| 10 |
-
lr_init: 0.00001
|
| 11 |
-
lr_end: 0.0001
|
| 12 |
-
warmup_steps: 2000
|
| 13 |
-
decay_steps: 40000
|
| 14 |
-
data:
|
| 15 |
-
max_eval_sample: 8
|
| 16 |
-
max_sec: 54
|
| 17 |
-
num_workers: 4
|
| 18 |
-
pad_val: 1024 # same with EOS in model
|
| 19 |
-
model:
|
| 20 |
-
vocab_size: 1025
|
| 21 |
-
phoneme_vocab_size: 732
|
| 22 |
-
embedding_dim: 512
|
| 23 |
-
hidden_dim: 512
|
| 24 |
-
head: 16
|
| 25 |
-
linear_units: 2048
|
| 26 |
-
n_layer: 24
|
| 27 |
-
dropout: 0
|
| 28 |
-
EOS: 1024
|
| 29 |
-
random_bert: 0
|
| 30 |
-
inference:
|
| 31 |
-
top_k: 15
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s1longer.yaml
DELETED
|
@@ -1,31 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 20
|
| 4 |
-
batch_size: 8
|
| 5 |
-
save_every_n_epoch: 1
|
| 6 |
-
precision: 16-mixed
|
| 7 |
-
gradient_clip: 1.0
|
| 8 |
-
optimizer:
|
| 9 |
-
lr: 0.01
|
| 10 |
-
lr_init: 0.00001
|
| 11 |
-
lr_end: 0.0001
|
| 12 |
-
warmup_steps: 2000
|
| 13 |
-
decay_steps: 40000
|
| 14 |
-
data:
|
| 15 |
-
max_eval_sample: 8
|
| 16 |
-
max_sec: 54
|
| 17 |
-
num_workers: 4
|
| 18 |
-
pad_val: 1024 # same with EOS in model
|
| 19 |
-
model:
|
| 20 |
-
vocab_size: 1025
|
| 21 |
-
phoneme_vocab_size: 512
|
| 22 |
-
embedding_dim: 512
|
| 23 |
-
hidden_dim: 512
|
| 24 |
-
head: 16
|
| 25 |
-
linear_units: 2048
|
| 26 |
-
n_layer: 24
|
| 27 |
-
dropout: 0
|
| 28 |
-
EOS: 1024
|
| 29 |
-
random_bert: 0
|
| 30 |
-
inference:
|
| 31 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s1mq.yaml
DELETED
|
@@ -1,77 +0,0 @@
|
|
| 1 |
-
train:
|
| 2 |
-
seed: 1234
|
| 3 |
-
epochs: 100
|
| 4 |
-
batch_size: 6
|
| 5 |
-
gradient_accumulation: 4
|
| 6 |
-
save_every_n_epoch: 1
|
| 7 |
-
precision: 32
|
| 8 |
-
gradient_clip: 1.0
|
| 9 |
-
optimizer:
|
| 10 |
-
lr: 0.01
|
| 11 |
-
lr_init: 0.00001
|
| 12 |
-
lr_end: 0.0001
|
| 13 |
-
warmup_steps: 2000
|
| 14 |
-
decay_steps: 40000
|
| 15 |
-
data:
|
| 16 |
-
max_eval_sample: 8
|
| 17 |
-
max_sec: 40
|
| 18 |
-
num_workers: 1
|
| 19 |
-
pad_val: 1024 # same with EOS in model
|
| 20 |
-
model:
|
| 21 |
-
saving_path: "ckpt/"
|
| 22 |
-
resume_checkpoint: null
|
| 23 |
-
vocoder_config_path: "quantizer/new_ckpt/config.json"
|
| 24 |
-
vocoder_ckpt_path: "quantizer/new_ckpt/g_00600000"
|
| 25 |
-
datadir: "/home/liweiche/GigaSpeech/wavs"
|
| 26 |
-
metapath: "/home/liweiche/GigaSpeech/train2.json"
|
| 27 |
-
val_metapath: "/home/liweiche/GigaSpeech/dev2.json"
|
| 28 |
-
sampledir: "logs/"
|
| 29 |
-
pretrained_path: null
|
| 30 |
-
lr: 0.0001
|
| 31 |
-
batch_size: 200.0
|
| 32 |
-
train_bucket_size: 8192
|
| 33 |
-
training_step: 800000
|
| 34 |
-
optim_flat_percent: 0.0
|
| 35 |
-
warmup_step: 50
|
| 36 |
-
adam_beta1: 0.9
|
| 37 |
-
adam_beta2: 0.98
|
| 38 |
-
ffd_size: 3072
|
| 39 |
-
hidden_size: 768
|
| 40 |
-
enc_nlayers: 6
|
| 41 |
-
dec_nlayers: 6
|
| 42 |
-
nheads: 12
|
| 43 |
-
ar_layer: 4
|
| 44 |
-
ar_ffd_size: 1024
|
| 45 |
-
ar_hidden_size: 256
|
| 46 |
-
ar_nheads: 4
|
| 47 |
-
aligner_softmax_temp: 1.0
|
| 48 |
-
layer_norm_eps: 0.00001
|
| 49 |
-
speaker_embed_dropout: 0.05
|
| 50 |
-
label_smoothing: 0.0
|
| 51 |
-
val_check_interval: 5000
|
| 52 |
-
check_val_every_n_epoch: 1
|
| 53 |
-
precision: "fp16"
|
| 54 |
-
nworkers: 16
|
| 55 |
-
distributed: true
|
| 56 |
-
accelerator: "ddp"
|
| 57 |
-
version: null
|
| 58 |
-
accumulate_grad_batches: 1
|
| 59 |
-
use_repetition_token: true
|
| 60 |
-
use_repetition_gating: false
|
| 61 |
-
repetition_penalty: 1.0
|
| 62 |
-
sampling_temperature: 1.0
|
| 63 |
-
top_k: -1
|
| 64 |
-
min_top_k: 3
|
| 65 |
-
top_p: 0.8
|
| 66 |
-
sample_num: 4
|
| 67 |
-
length_penalty_max_length: 15000
|
| 68 |
-
length_penalty_max_prob: 0.95
|
| 69 |
-
max_input_length: 2048
|
| 70 |
-
max_output_length: 2000
|
| 71 |
-
sample_rate: 16000
|
| 72 |
-
n_codes: 1024
|
| 73 |
-
n_cluster_groups: 1
|
| 74 |
-
phone_context_window: 4
|
| 75 |
-
phoneset_size: 1000
|
| 76 |
-
inference:
|
| 77 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/s2.json
DELETED
|
@@ -1,90 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"train": {
|
| 3 |
-
"log_interval": 100,
|
| 4 |
-
"eval_interval": 500,
|
| 5 |
-
"seed": 1234,
|
| 6 |
-
"epochs": 100,
|
| 7 |
-
"learning_rate": 0.0001,
|
| 8 |
-
"betas": [
|
| 9 |
-
0.8,
|
| 10 |
-
0.99
|
| 11 |
-
],
|
| 12 |
-
"eps": 1e-09,
|
| 13 |
-
"batch_size": 32,
|
| 14 |
-
"fp16_run": true,
|
| 15 |
-
"lr_decay": 0.999875,
|
| 16 |
-
"segment_size": 20480,
|
| 17 |
-
"init_lr_ratio": 1,
|
| 18 |
-
"warmup_epochs": 0,
|
| 19 |
-
"c_mel": 45,
|
| 20 |
-
"c_kl": 1.0,
|
| 21 |
-
"text_low_lr_rate": 0.4
|
| 22 |
-
},
|
| 23 |
-
"data": {
|
| 24 |
-
"max_wav_value": 32768.0,
|
| 25 |
-
"sampling_rate": 32000,
|
| 26 |
-
"filter_length": 2048,
|
| 27 |
-
"hop_length": 640,
|
| 28 |
-
"win_length": 2048,
|
| 29 |
-
"n_mel_channels": 128,
|
| 30 |
-
"mel_fmin": 0.0,
|
| 31 |
-
"mel_fmax": null,
|
| 32 |
-
"add_blank": true,
|
| 33 |
-
"n_speakers": 300,
|
| 34 |
-
"cleaned_text": true
|
| 35 |
-
},
|
| 36 |
-
"model": {
|
| 37 |
-
"inter_channels": 192,
|
| 38 |
-
"hidden_channels": 192,
|
| 39 |
-
"filter_channels": 768,
|
| 40 |
-
"n_heads": 2,
|
| 41 |
-
"n_layers": 6,
|
| 42 |
-
"kernel_size": 3,
|
| 43 |
-
"p_dropout": 0.1,
|
| 44 |
-
"resblock": "1",
|
| 45 |
-
"resblock_kernel_sizes": [
|
| 46 |
-
3,
|
| 47 |
-
7,
|
| 48 |
-
11
|
| 49 |
-
],
|
| 50 |
-
"resblock_dilation_sizes": [
|
| 51 |
-
[
|
| 52 |
-
1,
|
| 53 |
-
3,
|
| 54 |
-
5
|
| 55 |
-
],
|
| 56 |
-
[
|
| 57 |
-
1,
|
| 58 |
-
3,
|
| 59 |
-
5
|
| 60 |
-
],
|
| 61 |
-
[
|
| 62 |
-
1,
|
| 63 |
-
3,
|
| 64 |
-
5
|
| 65 |
-
]
|
| 66 |
-
],
|
| 67 |
-
"upsample_rates": [
|
| 68 |
-
10,
|
| 69 |
-
8,
|
| 70 |
-
2,
|
| 71 |
-
2,
|
| 72 |
-
2
|
| 73 |
-
],
|
| 74 |
-
"upsample_initial_channel": 512,
|
| 75 |
-
"upsample_kernel_sizes": [
|
| 76 |
-
16,
|
| 77 |
-
16,
|
| 78 |
-
8,
|
| 79 |
-
2,
|
| 80 |
-
2
|
| 81 |
-
],
|
| 82 |
-
"n_layers_q": 3,
|
| 83 |
-
"use_spectral_norm": false,
|
| 84 |
-
"gin_channels": 512,
|
| 85 |
-
"semantic_frame_rate": "25hz",
|
| 86 |
-
"freeze_quantizer": true
|
| 87 |
-
},
|
| 88 |
-
"s2_ckpt_dir": "logs/s2/big2k1",
|
| 89 |
-
"content_module": "cnhubert"
|
| 90 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
configs/train.yaml
DELETED
|
@@ -1,32 +0,0 @@
|
|
| 1 |
-
gpu:
|
| 2 |
-
n_card: 1
|
| 3 |
-
n_process_per_card: 2
|
| 4 |
-
io:
|
| 5 |
-
text_path: D:\RVC1006\GPT-SoVITS\GPT_SoVITS
|
| 6 |
-
save_every_n_epoch: 1
|
| 7 |
-
precision: 16-mixed
|
| 8 |
-
gradient_clip: 1.0
|
| 9 |
-
optimizer:
|
| 10 |
-
lr: 0.01
|
| 11 |
-
lr_init: 0.00001
|
| 12 |
-
lr_end: 0.0001
|
| 13 |
-
warmup_steps: 2000
|
| 14 |
-
decay_steps: 40000
|
| 15 |
-
data:
|
| 16 |
-
max_eval_sample: 8
|
| 17 |
-
max_sec: 54
|
| 18 |
-
num_workers: 1
|
| 19 |
-
pad_val: 1024 # same with EOS in model
|
| 20 |
-
model:
|
| 21 |
-
vocab_size: 1025
|
| 22 |
-
phoneme_vocab_size: 512
|
| 23 |
-
embedding_dim: 512
|
| 24 |
-
hidden_dim: 512
|
| 25 |
-
head: 16
|
| 26 |
-
linear_units: 2048
|
| 27 |
-
n_layer: 24
|
| 28 |
-
dropout: 0
|
| 29 |
-
EOS: 1024
|
| 30 |
-
random_bert: 0
|
| 31 |
-
inference:
|
| 32 |
-
top_k: 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|