Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -108,7 +108,7 @@ def get_model(name_model):
|
|
| 108 |
global models
|
| 109 |
if name_model in models:
|
| 110 |
return models[name_model]
|
| 111 |
-
models[name_model]=VitsModel.from_pretrained(name_model,token=token)
|
| 112 |
models[name_model].decoder.apply_weight_norm()
|
| 113 |
# torch.nn.utils.weight_norm(self.decoder.conv_pre)
|
| 114 |
# torch.nn.utils.weight_norm(self.decoder.conv_post)
|
|
@@ -130,7 +130,7 @@ def modelspeech(text=TXT,name_model="wasmdashai/vits-ar-sa-huba-v2",speaking_r
|
|
| 130 |
model=get_model(name_model)
|
| 131 |
model.speaking_rate=speaking_rate
|
| 132 |
with torch.no_grad():
|
| 133 |
-
wav=list(_inference_forward_stream(model,input_ids=inputs.input_ids
|
| 134 |
# with torch.no_grad():
|
| 135 |
# wav = model(input_ids=inputs["input_ids"].cuda()).waveform.cpu().numpy().reshape(-1)#.detach()
|
| 136 |
|
|
|
|
| 108 |
global models
|
| 109 |
if name_model in models:
|
| 110 |
return models[name_model]
|
| 111 |
+
models[name_model]=VitsModel.from_pretrained(name_model,token=token)
|
| 112 |
models[name_model].decoder.apply_weight_norm()
|
| 113 |
# torch.nn.utils.weight_norm(self.decoder.conv_pre)
|
| 114 |
# torch.nn.utils.weight_norm(self.decoder.conv_post)
|
|
|
|
| 130 |
model=get_model(name_model)
|
| 131 |
model.speaking_rate=speaking_rate
|
| 132 |
with torch.no_grad():
|
| 133 |
+
wav=list(_inference_forward_stream(model,input_ids=inputs.input_ids,attention_mask=inputs.attention_mask,speaker_embeddings= None,is_streaming=False))[0]
|
| 134 |
# with torch.no_grad():
|
| 135 |
# wav = model(input_ids=inputs["input_ids"].cuda()).waveform.cpu().numpy().reshape(-1)#.detach()
|
| 136 |
|