| Tortoise TTS AR model fine-tuned for German | |
| Trained on 3 speakers; 2 LibriVox readers, and Thorsten Mueller's dataset https://github.com/thorstenMueller/Thorsten-Voice | |
| ***THE NEWEST VERSIONS***: v# indicates the number of training sessions, #e is how many epochs. | |
| 9/5 training session uploaded | |
| Requires the tokenizer file placed in the tokenizers/ directory | |
| Voice latents are pre-computed in voices/ for some uploaded versions. Voice samples to recompute latents are uploaded. | |
| For use in MRQ Voice Cloning WebUI: | |
| Requires the tokenizer used in training, and code changes to disable text cleaners. At minimum, change english_cleaners to basic_cleaners. | |
| Code changes: | |
| modules\tortoise-tts\tortoise\utils\tokenizer.py | |
| Change Line 201: txt = english_cleaners(txt) and replace it | |
| with txt = basic_cleaners(txt) | |
| modules\tortoise-tts\build\lib\tortoise\utils\tokenizer.py | |
| Change Line 201: txt = english_cleaners(txt) and replace it | |
| with txt = basic_cleaners(txt) | |
| \modules\dlas\dlas\data\audio\paired_voice_audio_dataset.py | |
| Line 133: return text_to_sequence(txt, ['english_cleaners']) | |
| and replace it with: return text_to_sequence(txt, ['basic_cleaners']) | |
| modules\dlas\dlas\data\audio\voice_tokenizer.py | |
| Line 14: from dlas.models.audio.tts.tacotron2.text.cleaners import | |
| english_cleaners | |
| to: from dlas.models.audio.tts.tacotron2.text.cleaners import | |
| english_cleaners, basic_cleaners | |
| Line 85: txt = english_cleaners(txt) to txt = | |
| basic_cleaners(txt) | |
| Line 134: word = english_cleaners(word) to basic_cleaners(word) | |
| Copy and paste German text into the tokenizer tester on the utilities | |
| tab, and you should see it tokenized with all of the special | |
| characters, and no [UNK]. | |
| --- | |
| license: other | |
| language: | |
| - de | |
| --- |