| import os | |
| os.environ["COQUI_TOS_AGREED"] = "1" | |
| import pytest | |
| from tests.utils import wrap_test_forked | |
| from TTS.api import TTS | |
| def test_get_models(model_name): | |
| import torch | |
| from TTS.api import TTS | |
| # Get device | |
| device = "cuda" if torch.cuda.is_available() else "cpu" | |
| # Init TTS | |
| tts = TTS(model_name).to(device) | |
| # Run TTS | |
| # ❗ Since this model is multi-lingual voice cloning model, we must set the target speaker_wav and language | |
| # Text to speech list of amplitude values as output | |
| try: | |
| wav = tts.tts(text="Hello world!", speaker_wav="./models/male.wav", language="en") | |
| # Text to speech to a file | |
| tts.tts_to_file(text="Hello world!", speaker_wav="./models/male.wav", language="en", file_path="output.wav") | |
| except ValueError: | |
| wav = tts.tts(text="Hello world!", speaker_wav="./models/male.wav") | |
| # Text to speech to a file | |
| tts.tts_to_file(text="Hello world!", speaker_wav="./models/male.wav", file_path="output.wav") | |
| # files are located in e.g. /home/jon/.local/share/tts/tts_models--multilingual--multi-dataset--xtts_v1.1 | |
| # downloaded from e.g. https://coqui.gateway.scarf.sh/v0.6.1_models/tts_models--en--ljspeech--glow-tts.zip | |
| # all stored in https://h2o-release.s3.amazonaws.com/h2ogpt/tts_in_.local_share_tts.tgz |