aoxo commited on
Commit
dbc152d
·
verified ·
1 Parent(s): 67b4ec5

Update eduport_tts_mal.py

Browse files
Files changed (1) hide show
  1. eduport_tts_mal.py +20 -7
eduport_tts_mal.py CHANGED
@@ -72,26 +72,32 @@ class SpeechDataset(Dataset):
72
  def collate_fn(batch):
73
  audio_inputs, text_inputs = zip(*batch)
74
 
 
 
 
75
  # Pad audio inputs to the maximum length in the batch
76
  max_audio_len = max([audio.size(1) for audio in audio_inputs])
77
 
78
- # Create padded audio inputs
79
  audio_inputs_padded = []
80
  for audio in audio_inputs:
81
  if audio.size(1) < max_audio_len:
82
- # Create padding tensor with the same number of channels
83
- padding = torch.zeros(audio.size(0), max_audio_len - audio.size(1))
84
  padded_audio = torch.cat([audio, padding], dim=1)
85
  audio_inputs_padded.append(padded_audio)
86
  else:
87
  audio_inputs_padded.append(audio[:, :max_audio_len])
88
 
89
- # Convert to tensor
90
  audio_inputs_padded = torch.stack(audio_inputs_padded)
91
 
92
- # Pad text inputs to the longest transcript length
93
  max_text_len = max([text.size(0) for text in text_inputs])
94
- text_inputs_padded = torch.stack([torch.cat([text, torch.tensor([0] * (max_text_len - text.size(0)))], dim=0) if text.size(0) < max_text_len else text[:max_text_len] for text in text_inputs])
 
 
 
 
 
95
 
96
  return audio_inputs_padded, text_inputs_padded
97
 
@@ -142,6 +148,10 @@ class SpeechRecognitionModel(torch.nn.Module):
142
  self.decoder.gradient_checkpointing_enable()
143
 
144
  def forward(self, audio_input, text_input):
 
 
 
 
145
  # Extract encoder hidden states
146
  encoder_output = self.encoder(audio_input).last_hidden_state
147
 
@@ -224,7 +234,10 @@ def train_model(num_epochs=10, accumulation_steps=16):
224
  for batch_idx, (audio_input, text_input) in enumerate(train_progress):
225
 
226
  # Move tensors to device
227
- audio_input = audio_input.squeeze(1).to(device)
 
 
 
228
  text_input = text_input.to(device)
229
 
230
  # Use autocast for mixed precision training
 
72
  def collate_fn(batch):
73
  audio_inputs, text_inputs = zip(*batch)
74
 
75
+ # Ensure audio inputs are 3D
76
+ audio_inputs = [audio.squeeze(0) if audio.dim() == 3 else audio for audio in audio_inputs]
77
+
78
  # Pad audio inputs to the maximum length in the batch
79
  max_audio_len = max([audio.size(1) for audio in audio_inputs])
80
 
 
81
  audio_inputs_padded = []
82
  for audio in audio_inputs:
83
  if audio.size(1) < max_audio_len:
84
+ padding = torch.zeros(audio.size(0), max_audio_len - audio.size(1), device=audio.device)
 
85
  padded_audio = torch.cat([audio, padding], dim=1)
86
  audio_inputs_padded.append(padded_audio)
87
  else:
88
  audio_inputs_padded.append(audio[:, :max_audio_len])
89
 
90
+ # Stack audio inputs
91
  audio_inputs_padded = torch.stack(audio_inputs_padded)
92
 
93
+ # Pad text inputs
94
  max_text_len = max([text.size(0) for text in text_inputs])
95
+ text_inputs_padded = torch.stack([
96
+ torch.cat([text, torch.zeros(max_text_len - text.size(0), dtype=text.dtype)], dim=0)
97
+ if text.size(0) < max_text_len
98
+ else text[:max_text_len]
99
+ for text in text_inputs
100
+ ])
101
 
102
  return audio_inputs_padded, text_inputs_padded
103
 
 
148
  self.decoder.gradient_checkpointing_enable()
149
 
150
  def forward(self, audio_input, text_input):
151
+ # Ensure audio_input is 3D: [batch_size, channels, time]
152
+ if audio_input.dim() == 4:
153
+ audio_input = audio_input.squeeze(1) # Remove extra dimension
154
+
155
  # Extract encoder hidden states
156
  encoder_output = self.encoder(audio_input).last_hidden_state
157
 
 
234
  for batch_idx, (audio_input, text_input) in enumerate(train_progress):
235
 
236
  # Move tensors to device
237
+ audio_input = audio_input.squeeze.to(device)
238
+ # Squeeze or reshape if necessary
239
+ if audio_input.dim() == 4:
240
+ audio_input = audio_input.squeeze(1) # Remove extra singleton dimension
241
  text_input = text_input.to(device)
242
 
243
  # Use autocast for mixed precision training