shethjenil commited on
Commit
66bf596
·
verified ·
1 Parent(s): b02228a

Update modeling_vits.py

Browse files
Files changed (1) hide show
  1. modeling_vits.py +0 -1
modeling_vits.py CHANGED
@@ -26,7 +26,6 @@ class ModVitsModel(VitsModel):
26
 
27
  @torch.inference_mode()
28
  def forward(self, input_ids = None, attention_mask = None, speaker_id = None, output_attentions = None, output_hidden_states = None, return_dict = None, labels = None, **kwargs):
29
- # speaker_id = self.config.speaker_names.index(speaker_id) * len(self.config.emotion_names) + self.config.emotion_names.index(kwargs['emotion_id'])
30
  audio = super().forward(input_ids, attention_mask, speaker_id, output_attentions, output_hidden_states, return_dict, labels, **kwargs)
31
  B, T = audio.waveform.shape
32
  mask = torch.arange(T, device=audio.waveform.device).expand(B, T) < audio.sequence_lengths.unsqueeze(1)
 
26
 
27
  @torch.inference_mode()
28
  def forward(self, input_ids = None, attention_mask = None, speaker_id = None, output_attentions = None, output_hidden_states = None, return_dict = None, labels = None, **kwargs):
 
29
  audio = super().forward(input_ids, attention_mask, speaker_id, output_attentions, output_hidden_states, return_dict, labels, **kwargs)
30
  B, T = audio.waveform.shape
31
  mask = torch.arange(T, device=audio.waveform.device).expand(B, T) < audio.sequence_lengths.unsqueeze(1)