Update app.py
Browse files
app.py
CHANGED
|
@@ -20,25 +20,14 @@ def transcribe(audio):
|
|
| 20 |
result = whisper.decode(model, mel, options)
|
| 21 |
return result.text
|
| 22 |
|
| 23 |
-
input_audio = gr.Audio(
|
| 24 |
-
sources=["microphone"],
|
| 25 |
-
type="filepath",
|
| 26 |
-
|
| 27 |
-
waveform_options=gr.WaveformOptions(
|
| 28 |
-
waveform_color="#01C6FF",
|
| 29 |
-
waveform_progress_color="#0066B4",
|
| 30 |
-
skip_length=2,
|
| 31 |
-
show_controls=False,
|
| 32 |
-
),
|
| 33 |
-
)
|
| 34 |
|
| 35 |
gr.Interface(
|
| 36 |
title = 'OpenAI-Whisper Audio to Text Web UI',
|
| 37 |
fn=transcribe,
|
| 38 |
-
inputs
|
| 39 |
-
|
| 40 |
],
|
| 41 |
outputs=[
|
| 42 |
"textbox"
|
| 43 |
],
|
| 44 |
-
live=True).launch(
|
|
|
|
| 20 |
result = whisper.decode(model, mel, options)
|
| 21 |
return result.text
|
| 22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
gr.Interface(
|
| 25 |
title = 'OpenAI-Whisper Audio to Text Web UI',
|
| 26 |
fn=transcribe,
|
| 27 |
+
inputs=[
|
| 28 |
+
gr.inputs.Audio(source="microphone", type="filepath")
|
| 29 |
],
|
| 30 |
outputs=[
|
| 31 |
"textbox"
|
| 32 |
],
|
| 33 |
+
live=True).launch()
|