|
|
import torch |
|
|
from transformers import pipeline |
|
|
from cosyvoice.cli.cosyvoice import CosyVoice |
|
|
|
|
|
class EndpointHandler: |
|
|
def __init__(self, model): |
|
|
self.model = CosyVoice('path/to/model') |
|
|
self.device = 'cuda' if torch.cuda.is_available() else 'cpu' |
|
|
self.model.to(self.device) |
|
|
|
|
|
def __call__(self, data): |
|
|
inputs = data.get('inputs', 'Default text') |
|
|
parameters = data.get('parameters', {}) |
|
|
language = parameters.get('language', 'en') |
|
|
duration = parameters.get('duration', 30) |
|
|
|
|
|
|
|
|
audio = self.model.generate(inputs, language=language, singing_mode=True) |
|
|
|
|
|
|
|
|
import io |
|
|
import base64 |
|
|
buffer = io.BytesIO() |
|
|
torch.save(audio, buffer) |
|
|
audio_b64 = base64.b64encode(buffer.getvalue()).decode() |
|
|
|
|
|
return {'audio': audio_b64} |
|
|
|
|
|
|
|
|
def query_endpoint(payload): |
|
|
handler = EndpointHandler(None) |
|
|
return handler(payload) |