import torch from torch.utils.data import Dataset, DataLoader from src.text_tokenizer import TextTokenizer class MyDataset(Dataset): def __init__(self, hf_dataset, train_type): self.hf_dataset = hf_dataset self.text_tokenizer = TextTokenizer() assert train_type in ['pretrain', 'sft'] self.train_type = train_type def __len__(self): return len(self.hf_dataset) def __getitem__(self, idx): data = self.hf_dataset[idx] audio_token_list = data['audio_token'] text = data['normalized_text'] # text = data['text_normalized'] # text = "["+data['speaker_id']+ "]"+data['text_normalized'] if(self.train_type == 'sft'): text_token_list = self.text_tokenizer.tokenize(text) return text_token_list, audio_token_list else: return audio_token_list