Spaces:
Sleeping
Sleeping
Update conver.py
Browse files
conver.py
CHANGED
|
@@ -140,8 +140,8 @@ class URLToAudioConverter:
|
|
| 140 |
self.llm_out = conversation_json
|
| 141 |
audio_files, folder_name = await self.text_to_speech(conversation_json, voice_1, voice_2)
|
| 142 |
combined_audio = self.combine_audio_files(audio_files)
|
| 143 |
-
music_path = "
|
| 144 |
-
tags_paths = ["
|
| 145 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 146 |
final_output = os.path.join(folder_name, "combined_output_with_music.mp3")
|
| 147 |
final_audio.export(final_output, format="mp3")
|
|
@@ -156,8 +156,8 @@ class URLToAudioConverter:
|
|
| 156 |
)
|
| 157 |
audio_files, folder_name = await self.text_to_speech(conversation_json, voice_1, voice_2)
|
| 158 |
combined_audio = self.combine_audio_files(audio_files)
|
| 159 |
-
music_path = "
|
| 160 |
-
tags_paths = ["
|
| 161 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 162 |
final_output = os.path.join(folder_name, "combined_output_with_music.mp3")
|
| 163 |
final_audio.export(final_output, format="mp3")
|
|
@@ -166,16 +166,18 @@ class URLToAudioConverter:
|
|
| 166 |
return final_output, conversation_text
|
| 167 |
|
| 168 |
async def generate_podcast(self, conversation, voice_1, voice_2):
|
| 169 |
-
folder_name = await self.text_to_speech(conversation, voice_1, voice_2)
|
| 170 |
combined_audio = self.combine_audio_files(audio_files)
|
| 171 |
-
music_path = "musica.mp3"
|
| 172 |
-
tags_paths = ["tag.mp3", "tag2.mp3"]
|
| 173 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 174 |
output_file = os.path.join(folder_name, "raw_podcast_with_music.mp3")
|
| 175 |
final_audio.export(output_file, format="mp3")
|
| 176 |
-
|
| 177 |
-
# Limpieza de archivos temporales
|
| 178 |
for f in audio_files:
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
|
|
|
|
|
|
|
|
|
|
|
| 140 |
self.llm_out = conversation_json
|
| 141 |
audio_files, folder_name = await self.text_to_speech(conversation_json, voice_1, voice_2)
|
| 142 |
combined_audio = self.combine_audio_files(audio_files)
|
| 143 |
+
music_path = "musica.mp3"
|
| 144 |
+
tags_paths = ["tag.mp3", "tag2.mp3"]
|
| 145 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 146 |
final_output = os.path.join(folder_name, "combined_output_with_music.mp3")
|
| 147 |
final_audio.export(final_output, format="mp3")
|
|
|
|
| 156 |
)
|
| 157 |
audio_files, folder_name = await self.text_to_speech(conversation_json, voice_1, voice_2)
|
| 158 |
combined_audio = self.combine_audio_files(audio_files)
|
| 159 |
+
music_path = "musica.mp3"
|
| 160 |
+
tags_paths = ["tag.mp3", "tag2.mp3"]
|
| 161 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 162 |
final_output = os.path.join(folder_name, "combined_output_with_music.mp3")
|
| 163 |
final_audio.export(final_output, format="mp3")
|
|
|
|
| 166 |
return final_output, conversation_text
|
| 167 |
|
| 168 |
async def generate_podcast(self, conversation, voice_1, voice_2):
|
| 169 |
+
audio_files, folder_name = await self.text_to_speech(conversation, voice_1, voice_2)
|
| 170 |
combined_audio = self.combine_audio_files(audio_files)
|
| 171 |
+
music_path = "musica.mp3"
|
| 172 |
+
tags_paths = ["tag.mp3", "tag2.mp3"]
|
| 173 |
final_audio = self.add_background_music_and_tags(combined_audio, music_path, tags_paths)
|
| 174 |
output_file = os.path.join(folder_name, "raw_podcast_with_music.mp3")
|
| 175 |
final_audio.export(output_file, format="mp3")
|
| 176 |
+
|
|
|
|
| 177 |
for f in audio_files:
|
| 178 |
+
os.remove(f)
|
| 179 |
+
|
| 180 |
+
conversation_text = "\n".join(
|
| 181 |
+
f"{turn['speaker']}: {turn['text']}" for turn in conversation["conversation"]
|
| 182 |
+
)
|
| 183 |
+
return conversation_text, output_file
|