import os import torch import torchaudio from tqdm import tqdm from chatterbox.tts_turbo import ChatterboxTurboTTS DEVICE = "cuda" if torch.cuda.is_available() else "cpu" BASE_DIR = "/home/cloud/StyleTTS2-fine-tuning" OUTPUT_DIR = os.path.join(BASE_DIR, "Data") REFERENCE_AUDIO_PATH = os.path.join(OUTPUT_DIR, "reference_wavs/british_accent_audio.wav") INPUT_TEXT_FILE = os.path.join(OUTPUT_DIR, "source_text_final.txt") TRAIN_LIST_PATH = os.path.join(OUTPUT_DIR, "train_list_new.txt") WAVS_DIR = os.path.join(OUTPUT_DIR, "wavs") TARGET_SAMPLE_RATE = 24000 os.makedirs(WAVS_DIR, exist_ok=True) model = ChatterboxTurboTTS.from_pretrained(device=DEVICE) def get_sentences(text_path): if not os.path.exists(text_path): return [] with open(text_path, 'r', encoding='utf-8') as f: lines = f.readlines() valid_sentences = [] for line in lines: cleaned = line.strip() if cleaned: valid_sentences.append(cleaned) return valid_sentences def get_completed_indices(): if not os.path.exists(TRAIN_LIST_PATH): return set() completed = set() with open(TRAIN_LIST_PATH, "r", encoding="utf-8") as f: for line in f: parts = line.strip().split("|") if parts and len(parts) >= 1: filename = parts[0] try: number_part = filename.split("_")[1].split(".")[0] completed.add(int(number_part)) except: continue return completed def generate_dataset(): sentences = get_sentences(INPUT_TEXT_FILE) completed_indices = get_completed_indices() print(f"Total sentences: {len(sentences)}") print(f"Already done: {len(completed_indices)}") resampler = None if model.sr != TARGET_SAMPLE_RATE: resampler = torchaudio.transforms.Resample(orig_freq=model.sr, new_freq=TARGET_SAMPLE_RATE).to(DEVICE) for i, sentence in enumerate(tqdm(sentences)): if (i + 1) in completed_indices: continue filename = f"file_{i+1:04d}.wav" filepath = os.path.join(WAVS_DIR, filename) print(f"Generating {filename}...") try: with torch.inference_mode(): wav_tensor = model.generate( sentence, audio_prompt_path=REFERENCE_AUDIO_PATH ) if wav_tensor.dim() == 1: wav_tensor = wav_tensor.unsqueeze(0) if wav_tensor.shape[0] > 1: wav_tensor = wav_tensor.mean(dim=0, keepdim=True) wav_tensor = wav_tensor.cpu() if resampler: wav_tensor = resampler(wav_tensor) torchaudio.save(filepath, wav_tensor, TARGET_SAMPLE_RATE) with open(TRAIN_LIST_PATH, "a", encoding="utf-8") as f: f.write(f"{filename}|{sentence}|0\n") del wav_tensor torch.cuda.empty_cache() if (i + 1) % 50 == 0: torch.cuda.synchronize() except Exception as e: print(f"Error at sample {i+1}: {e}") continue if __name__ == "__main__": generate_dataset()