Update app.py
Browse files
app.py
CHANGED
|
@@ -1,18 +1,17 @@
|
|
| 1 |
from elevenlabs import VoiceSettings
|
| 2 |
from elevenlabs.client import ElevenLabs
|
| 3 |
from transformers import M2M100ForConditionalGeneration, M2M100Tokenizer
|
| 4 |
-
import whisper
|
| 5 |
from ai71 import AI71
|
| 6 |
from datetime import datetime
|
| 7 |
import os
|
| 8 |
import time
|
| 9 |
from pydub import AudioSegment
|
| 10 |
-
# from IPython.display import Audio, display, Video, HTML
|
| 11 |
-
# import assemblyai as aai
|
| 12 |
from base64 import b64encode
|
| 13 |
import gradio as gr
|
| 14 |
import concurrent.futures
|
| 15 |
-
import
|
|
|
|
|
|
|
| 16 |
|
| 17 |
# aai.settings.api_key = "d5b107f34d534b4ebdfbd869f8408f92"
|
| 18 |
# transcriber = aai.Transcriber()
|
|
@@ -163,8 +162,9 @@ def summarize(meeting_texts=meeting_texts):
|
|
| 163 |
|
| 164 |
# Placeholder function for speech to text conversion
|
| 165 |
def speech_to_text(video):
|
|
|
|
| 166 |
print('Started transcribing')
|
| 167 |
-
audio = AudioSegment.from_file(video
|
| 168 |
audio.export('temp.wav', format="wav")
|
| 169 |
|
| 170 |
# transcript = transcriber.transcribe(video).text
|
|
|
|
| 1 |
from elevenlabs import VoiceSettings
|
| 2 |
from elevenlabs.client import ElevenLabs
|
| 3 |
from transformers import M2M100ForConditionalGeneration, M2M100Tokenizer
|
|
|
|
| 4 |
from ai71 import AI71
|
| 5 |
from datetime import datetime
|
| 6 |
import os
|
| 7 |
import time
|
| 8 |
from pydub import AudioSegment
|
|
|
|
|
|
|
| 9 |
from base64 import b64encode
|
| 10 |
import gradio as gr
|
| 11 |
import concurrent.futures
|
| 12 |
+
# from IPython.display import Audio, display, Video, HTML
|
| 13 |
+
# import assemblyai as aai
|
| 14 |
+
# import whisper
|
| 15 |
|
| 16 |
# aai.settings.api_key = "d5b107f34d534b4ebdfbd869f8408f92"
|
| 17 |
# transcriber = aai.Transcriber()
|
|
|
|
| 162 |
|
| 163 |
# Placeholder function for speech to text conversion
|
| 164 |
def speech_to_text(video):
|
| 165 |
+
print(video, type(video))
|
| 166 |
print('Started transcribing')
|
| 167 |
+
audio = AudioSegment.from_file(video)
|
| 168 |
audio.export('temp.wav', format="wav")
|
| 169 |
|
| 170 |
# transcript = transcriber.transcribe(video).text
|