ayymen commited on
Commit
ae4b5f4
·
verified ·
1 Parent(s): 0f7a61e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/trainer_0_log.txt filter=lfs diff=lfs merge=lfs -text
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ff160ff4136e27a514dec50d97d2123e02e9155cf90f66d9f94badca8fa55e2
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model_1111040.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c79be5d192061e2fef0d5b6bc35fc6147b448a8e9378fc68cef8875d4d096832
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model_1249920.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:454c7f761c14d970575621c1daca54ecf7e5ca6190c563dc0accc89fbaa465ca
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model_1631840.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e495d27884f790bbcf1d5ecd1e0e2b864a929cda8ee118d6e037b336bf2fee10
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model_1666560.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:900fd532b8796404c85ed249290f64fd22d072cb81115472f939aaf791e72f26
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/best_model_1736000.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ff160ff4136e27a514dec50d97d2123e02e9155cf90f66d9f94badca8fa55e2
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/checkpoint_1730000.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d90787ab7f398ef5425cc38a71f44b6e98105ca8c67f263284be8f11260c998
3
+ size 5648645125
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/config.json ADDED
@@ -0,0 +1,215 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "xtts_hausa",
3
+ "logger_uri": null,
4
+ "run_name": "GPT_XTTS_HAUSA_FT",
5
+ "project_name": "XTTS_trainer",
6
+ "run_description": [
7
+ "\n GPT XTTS training\n "
8
+ ],
9
+ "print_step": 50,
10
+ "plot_step": 100,
11
+ "model_param_stats": false,
12
+ "wandb_entity": null,
13
+ "dashboard_logger": "tensorboard",
14
+ "save_on_interrupt": true,
15
+ "log_model_step": 100,
16
+ "save_step": 10000,
17
+ "save_n_checkpoints": 10,
18
+ "save_checkpoints": true,
19
+ "save_all_best": true,
20
+ "save_best_after": 0,
21
+ "target_loss": null,
22
+ "print_eval": false,
23
+ "test_delay_epochs": 0,
24
+ "run_eval": true,
25
+ "run_eval_steps": null,
26
+ "distributed_backend": "nccl",
27
+ "distributed_url": "tcp://localhost:54321",
28
+ "mixed_precision": true,
29
+ "precision": "bf16",
30
+ "epochs": 1000,
31
+ "batch_size": 1,
32
+ "eval_batch_size": 1,
33
+ "grad_clip": 0.0,
34
+ "scheduler_after_epoch": true,
35
+ "lr": 5e-06,
36
+ "optimizer": "AdamW",
37
+ "optimizer_params": {
38
+ "betas": [
39
+ 0.9,
40
+ 0.96
41
+ ],
42
+ "eps": 1e-08,
43
+ "weight_decay": 0.01
44
+ },
45
+ "lr_scheduler": "MultiStepLR",
46
+ "lr_scheduler_params": {
47
+ "milestones": [
48
+ 900000,
49
+ 2700000,
50
+ 5400000
51
+ ],
52
+ "gamma": 0.5,
53
+ "last_epoch": -1
54
+ },
55
+ "use_grad_scaler": false,
56
+ "allow_tf32": false,
57
+ "cudnn_enable": true,
58
+ "cudnn_deterministic": false,
59
+ "cudnn_benchmark": false,
60
+ "training_seed": 54321,
61
+ "model": "xtts",
62
+ "num_loader_workers": 8,
63
+ "num_eval_loader_workers": 0,
64
+ "use_noise_augment": false,
65
+ "audio": {
66
+ "sample_rate": 22050,
67
+ "output_sample_rate": 24000,
68
+ "dvae_sample_rate": 22050
69
+ },
70
+ "use_phonemes": false,
71
+ "phonemizer": null,
72
+ "phoneme_language": null,
73
+ "compute_input_seq_cache": false,
74
+ "text_cleaner": null,
75
+ "enable_eos_bos_chars": false,
76
+ "test_sentences_file": "",
77
+ "phoneme_cache_path": null,
78
+ "characters": null,
79
+ "add_blank": false,
80
+ "batch_group_size": 0,
81
+ "loss_masking": null,
82
+ "min_audio_len": 1,
83
+ "max_audio_len": Infinity,
84
+ "min_text_len": 1,
85
+ "max_text_len": Infinity,
86
+ "compute_f0": false,
87
+ "compute_energy": false,
88
+ "compute_linear_spec": false,
89
+ "precompute_num_workers": 0,
90
+ "start_by_longest": false,
91
+ "shuffle": false,
92
+ "drop_last": false,
93
+ "datasets": [
94
+ {
95
+ "formatter": "",
96
+ "dataset_name": "",
97
+ "path": "",
98
+ "meta_file_train": "",
99
+ "ignored_speakers": null,
100
+ "language": "",
101
+ "phonemizer": "",
102
+ "meta_file_val": "",
103
+ "meta_file_attn_mask": ""
104
+ }
105
+ ],
106
+ "test_sentences": [
107
+ {
108
+ "text": "Umarnai don zaman tsarki.",
109
+ "speaker_wav": [
110
+ "/app/data/clips/JOS_005_008.wav"
111
+ ],
112
+ "language": "ha"
113
+ },
114
+ {
115
+ "text": "wanda kuma ya fa\u0257a mana \u0199aunar da kuke yi cikin Ruhu.",
116
+ "speaker_wav": [
117
+ "/app/data/clips/JOS_005_008.wav"
118
+ ],
119
+ "language": "ha"
120
+ },
121
+ {
122
+ "text": "Gama mun ji labarin bangaskiyarku a cikin Yesu Kiristi da kuma \u0199aunar da kuke yi saboda dukan tsarkaka.",
123
+ "speaker_wav": [
124
+ "/app/data/clips/JOS_005_008.wav"
125
+ ],
126
+ "language": "ha"
127
+ }
128
+ ],
129
+ "eval_split_max_size": null,
130
+ "eval_split_size": 0.01,
131
+ "use_speaker_weighted_sampler": false,
132
+ "speaker_weighted_sampler_alpha": 1.0,
133
+ "use_language_weighted_sampler": false,
134
+ "language_weighted_sampler_alpha": 1.0,
135
+ "use_length_weighted_sampler": false,
136
+ "length_weighted_sampler_alpha": 1.0,
137
+ "model_args": {
138
+ "gpt_batch_size": 1,
139
+ "enable_redaction": false,
140
+ "kv_cache": true,
141
+ "gpt_checkpoint": "",
142
+ "clvp_checkpoint": null,
143
+ "decoder_checkpoint": null,
144
+ "num_chars": 255,
145
+ "tokenizer_file": "xtts_hausa/XTTS_v2.0_original_model_files/vocab.json",
146
+ "gpt_max_audio_tokens": 605,
147
+ "gpt_max_text_tokens": 402,
148
+ "gpt_max_prompt_tokens": 70,
149
+ "gpt_layers": 30,
150
+ "gpt_n_model_channels": 1024,
151
+ "gpt_n_heads": 16,
152
+ "gpt_number_text_tokens": 8337,
153
+ "gpt_start_text_token": 261,
154
+ "gpt_stop_text_token": 0,
155
+ "gpt_num_audio_tokens": 1026,
156
+ "gpt_start_audio_token": 1024,
157
+ "gpt_stop_audio_token": 1025,
158
+ "gpt_code_stride_len": 1024,
159
+ "gpt_use_masking_gt_prompt_approach": true,
160
+ "gpt_use_perceiver_resampler": true,
161
+ "input_sample_rate": 22050,
162
+ "output_sample_rate": 24000,
163
+ "output_hop_length": 256,
164
+ "decoder_input_dim": 1024,
165
+ "d_vector_dim": 512,
166
+ "cond_d_vector_in_each_upsampling_layer": true,
167
+ "duration_const": 102400,
168
+ "min_conditioning_length": 11025,
169
+ "max_conditioning_length": 132300,
170
+ "gpt_loss_text_ce_weight": 0.01,
171
+ "gpt_loss_mel_ce_weight": 1.0,
172
+ "debug_loading_failures": true,
173
+ "max_wav_length": 264600,
174
+ "max_text_length": 300,
175
+ "mel_norm_file": "xtts_hausa/XTTS_v2.0_original_model_files/mel_stats.pth",
176
+ "dvae_checkpoint": "xtts_hausa/XTTS_v2.0_original_model_files/dvae.pth",
177
+ "xtts_checkpoint": "xtts_hausa/XTTS_v2.0_original_model_files/model.pth",
178
+ "vocoder": ""
179
+ },
180
+ "model_dir": null,
181
+ "languages": [
182
+ "en",
183
+ "es",
184
+ "fr",
185
+ "de",
186
+ "it",
187
+ "pt",
188
+ "pl",
189
+ "tr",
190
+ "ru",
191
+ "nl",
192
+ "cs",
193
+ "ar",
194
+ "zh-cn",
195
+ "hu",
196
+ "ko",
197
+ "ja",
198
+ "hi",
199
+ "ha"
200
+ ],
201
+ "temperature": 0.75,
202
+ "length_penalty": 1.0,
203
+ "repetition_penalty": 5.0,
204
+ "top_k": 50,
205
+ "top_p": 0.85,
206
+ "num_gpt_outputs": 1,
207
+ "gpt_cond_len": 30,
208
+ "gpt_cond_chunk_len": 4,
209
+ "max_ref_len": 30,
210
+ "sound_norm_refs": false,
211
+ "optimizer_wd_only_on_weights": true,
212
+ "weighted_loss_attrs": {},
213
+ "weighted_loss_multipliers": {},
214
+ "github_branch": "* main"
215
+ }
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/events.out.tfevents.1736419242.d90d290394eb.1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29e901ccff7f95f8d78877553043d06607b7aaef93b3e6c543ea0a5aef82d37d
3
+ size 36454819
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/train_gpt_xtts.py ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+
3
+ from trainer import Trainer, TrainerArgs
4
+
5
+ from TTS.config.shared_configs import BaseDatasetConfig
6
+ from TTS.tts.datasets import load_tts_samples
7
+ from TTS.tts.layers.xtts.trainer.gpt_trainer import GPTArgs, GPTTrainer, GPTTrainerConfig, XttsAudioConfig
8
+ from TTS.utils.manage import ModelManager
9
+ from math import ceil
10
+
11
+ # Logging parameters
12
+ RUN_NAME = "GPT_XTTS_HAUSA_FT"
13
+ PROJECT_NAME = "XTTS_trainer"
14
+ DASHBOARD_LOGGER = "tensorboard"
15
+ LOGGER_URI = None
16
+
17
+ # Set here the path that the checkpoints will be saved.
18
+ OUT_PATH = "xtts_hausa"
19
+
20
+ # Training Parameters
21
+ OPTIMIZER_WD_ONLY_ON_WEIGHTS = True # for multi-gpu training please make it False
22
+ START_WITH_EVAL = True # if True it will star with evaluation
23
+ BATCH_SIZE = 1 # set here the batch size
24
+ GRAD_ACUMM_STEPS = ceil(252 / BATCH_SIZE) # set here the grad accumulation steps
25
+ # Note: we recommend that BATCH_SIZE * GRAD_ACUMM_STEPS need to be at least 252 for more efficient training. You can increase/decrease BATCH_SIZE but then set GRAD_ACUMM_STEPS accordingly.
26
+
27
+ # Define here the dataset that you want to use for the fine-tuning on.
28
+ config_dataset = BaseDatasetConfig(
29
+ formatter="coqui",
30
+ dataset_name="ft_dataset",
31
+ path="data/",
32
+ meta_file_train="manifest_train.csv",
33
+ meta_file_val="manifest_dev.csv",
34
+ language="ha",
35
+ )
36
+
37
+ # Add here the configs of the datasets
38
+ DATASETS_CONFIG_LIST = [config_dataset]
39
+
40
+ # Define the path where XTTS v2.0.1 files will be downloaded
41
+ CHECKPOINTS_OUT_PATH = os.path.join(OUT_PATH, "XTTS_v2.0_original_model_files/")
42
+ os.makedirs(CHECKPOINTS_OUT_PATH, exist_ok=True)
43
+
44
+
45
+ # DVAE files
46
+ DVAE_CHECKPOINT_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/dvae.pth"
47
+ MEL_NORM_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/mel_stats.pth"
48
+
49
+ # Set the path to the downloaded files
50
+ DVAE_CHECKPOINT = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(DVAE_CHECKPOINT_LINK))
51
+ MEL_NORM_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(MEL_NORM_LINK))
52
+
53
+ # download DVAE files if needed
54
+ if not os.path.isfile(DVAE_CHECKPOINT) or not os.path.isfile(MEL_NORM_FILE):
55
+ print(" > Downloading DVAE files!")
56
+ ModelManager._download_model_files([MEL_NORM_LINK, DVAE_CHECKPOINT_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True)
57
+
58
+
59
+ # Download XTTS v2.0 checkpoint if needed
60
+ TOKENIZER_FILE_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/vocab.json"
61
+ XTTS_CHECKPOINT_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/model.pth"
62
+ XTTS_CONFIG_LINK = "https://coqui.gateway.scarf.sh/hf-coqui/XTTS-v2/main/config.json"
63
+
64
+ # XTTS transfer learning parameters: You we need to provide the paths of XTTS model checkpoint that you want to do the fine tuning.
65
+ TOKENIZER_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(TOKENIZER_FILE_LINK)) # vocab.json file
66
+ XTTS_CHECKPOINT = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(XTTS_CHECKPOINT_LINK)) # model.pth file
67
+ XTTS_CONFIG_FILE = os.path.join(CHECKPOINTS_OUT_PATH, os.path.basename(XTTS_CONFIG_LINK)) # config.json file
68
+
69
+ # download XTTS v2.0 files if needed
70
+ if not os.path.isfile(TOKENIZER_FILE):
71
+ print(" > Downloading XTTS v2.0 tokenizer!")
72
+ ModelManager._download_model_files(
73
+ [TOKENIZER_FILE_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
74
+ )
75
+ if not os.path.isfile(XTTS_CHECKPOINT):
76
+ print(" > Downloading XTTS v2.0 checkpoint!")
77
+ ModelManager._download_model_files(
78
+ [XTTS_CHECKPOINT_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
79
+ )
80
+ if not os.path.isfile(XTTS_CONFIG_FILE):
81
+ print(" > Downloading XTTS v2.0 config!")
82
+ ModelManager._download_model_files(
83
+ [XTTS_CONFIG_LINK], CHECKPOINTS_OUT_PATH, progress_bar=True
84
+ )
85
+
86
+ # load training samples
87
+ train_samples, eval_samples = load_tts_samples(
88
+ DATASETS_CONFIG_LIST,
89
+ eval_split=True,
90
+ )
91
+ print(f"Train samples: {len(train_samples)}")
92
+ print(f"Eval samples: {len(eval_samples)}")
93
+
94
+ # get the longest text audio file to use as speaker reference
95
+ samples_len = [len(item["text"].split(" ")) for item in train_samples]
96
+ longest_text_idx = samples_len.index(max(samples_len))
97
+ SPEAKER_REFERENCE = [train_samples[longest_text_idx]["audio_file"]] # speaker reference to be used in training test sentences
98
+ LANGUAGE = config_dataset.language
99
+
100
+
101
+ def main():
102
+ # init args and config
103
+ model_args = GPTArgs(
104
+ max_conditioning_length=132300, # 6 secs
105
+ min_conditioning_length=11025, # 0.5 secs
106
+ debug_loading_failures=True,
107
+ max_wav_length=12*22050, # 12 secs
108
+ max_text_length=300,
109
+ mel_norm_file=MEL_NORM_FILE,
110
+ dvae_checkpoint=DVAE_CHECKPOINT,
111
+ xtts_checkpoint=XTTS_CHECKPOINT, # checkpoint path of the model that you want to fine-tune
112
+ tokenizer_file=TOKENIZER_FILE,
113
+ gpt_num_audio_tokens=1026,
114
+ gpt_start_audio_token=1024,
115
+ gpt_stop_audio_token=1025,
116
+ gpt_use_masking_gt_prompt_approach=True,
117
+ gpt_use_perceiver_resampler=True,
118
+ )
119
+ # define audio config
120
+ audio_config = XttsAudioConfig(sample_rate=22050, dvae_sample_rate=22050, output_sample_rate=24000)
121
+ # training parameters config
122
+ config = GPTTrainerConfig()
123
+
124
+ config.load_json(XTTS_CONFIG_FILE)
125
+
126
+ config.mixed_precision = True
127
+ config.precision = "bf16"
128
+ config.epochs = 1000
129
+ config.output_path = OUT_PATH
130
+ config.model_args = model_args
131
+ config.run_name = RUN_NAME
132
+ config.project_name = PROJECT_NAME
133
+ config.run_description = """
134
+ GPT XTTS training
135
+ """,
136
+ config.dashboard_logger = DASHBOARD_LOGGER
137
+ config.logger_uri = LOGGER_URI
138
+ config.audio = audio_config
139
+ config.batch_size = BATCH_SIZE
140
+ config.eval_batch_size = BATCH_SIZE
141
+ config.num_loader_workers = 8
142
+ config.print_step = 50
143
+ config.plot_step = 100
144
+ config.log_model_step = 100
145
+ config.save_step = 10000
146
+ config.save_n_checkpoints = 10
147
+ config.save_checkpoints = True
148
+ config.save_all_best = True
149
+ config.save_best_after = 0
150
+ config.print_eval = False
151
+ # Optimizer values like tortoise, pytorch implementation with modifications to not apply WD to non-weight parameters.
152
+ config.optimizer = "AdamW"
153
+ config.optimizer_wd_only_on_weights = OPTIMIZER_WD_ONLY_ON_WEIGHTS
154
+ config.optimizer_params = {"betas": [0.9, 0.96], "eps": 1e-8, "weight_decay": 1e-2}
155
+ config.lr = 5e-06 # learning rate
156
+ config.lr_scheduler = "MultiStepLR"
157
+ config.lr_scheduler_params = {"milestones": [50000 * 18, 150000 * 18, 300000 * 18], "gamma": 0.5, "last_epoch": -1}
158
+ config.test_sentences=[
159
+ {
160
+ "text": "Umarnai don zaman tsarki.",
161
+ "speaker_wav": SPEAKER_REFERENCE,
162
+ "language": LANGUAGE,
163
+ },
164
+ {
165
+ "text": "wanda kuma ya faɗa mana ƙaunar da kuke yi cikin Ruhu.",
166
+ "speaker_wav": SPEAKER_REFERENCE,
167
+ "language": LANGUAGE,
168
+ },
169
+ {
170
+ "text": "Gama mun ji labarin bangaskiyarku a cikin Yesu Kiristi da kuma ƙaunar da kuke yi saboda dukan tsarkaka.",
171
+ "speaker_wav": SPEAKER_REFERENCE,
172
+ "language": LANGUAGE,
173
+ }
174
+ ]
175
+
176
+ # init the model from config
177
+ model = GPTTrainer.init_from_config(config)
178
+
179
+ # init the trainer and 🚀
180
+ trainer = Trainer(
181
+ TrainerArgs(
182
+ restore_path=None, # xtts checkpoint is restored via xtts_checkpoint key so no need of restore it using Trainer restore_path parameter
183
+ skip_train_epoch=False,
184
+ start_with_eval=START_WITH_EVAL,
185
+ grad_accum_steps=GRAD_ACUMM_STEPS,
186
+ ),
187
+ config,
188
+ output_path=OUT_PATH,
189
+ model=model,
190
+ train_samples=train_samples,
191
+ eval_samples=eval_samples,
192
+ )
193
+ trainer.fit()
194
+
195
+
196
+ if __name__ == "__main__":
197
+ main()
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/trainer_0_log.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9eb0a311b631bf1a307122d04ad4c1d78c4fb76248ed21fcea3f37fc92ffeeda
3
+ size 13083418
GPT_XTTS_HAUSA_FT-January-09-2025_10+40AM-8e59ec3/vocab.json ADDED
The diff for this file is too large to render. See raw diff