F5-TTS-mongolian / README.md
mendeeb's picture
Create README.md
43bb3db verified

F5-TTS trained on Mongolian speech dataset. Epoch was around 33 when I stopped it and (I initially set 100 but too long) and updates were 51000.

Parameters that I know of:

Base model: F5TTS Base Epochs: 100 Learning rage: 0.000075 Max Gradient Norm: 1 Warmup updates: 57 Batch Size Type: frame Batch size per gpu: 1600 (rtx 3080ti) grad_acc_steps = 1 max_samples = 64 precision: fp16 logger: wandb