TTS
Collection
4 items • Updated
How to use BricksDisplay/ellie-Bert-VITS2 with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("text-to-speech", model="BricksDisplay/ellie-Bert-VITS2", trust_remote_code=True) # Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("BricksDisplay/ellie-Bert-VITS2", trust_remote_code=True, dtype="auto")Taiwan accent TTS model from JackEllie.
Using this checkpoint from Hugging Face Transformers:
from transformers import AutoModel, AutoProcessor
from scipy.io.wavfile import write
import torch
model_name = "BricksDisplay/ellie-Bert-VITS2"
model = AutoModel.from_pretrained(model_name, trust_remote_code=True)
processor = AutoProcessor.from_pretrained(model_name, trust_remote_code=True)
with torch.no_grad():
inputs = processor("ä½ å¥½", language="zh", return_tensors="pt")
result = model(**inputs)
result = result["waveform"]
write("output.wav", model.config.sampling_rate, result[0].numpy())