Update app.py
Browse files
app.py
CHANGED
|
@@ -94,6 +94,7 @@ def generate_image(api_key, text):
|
|
| 94 |
return None
|
| 95 |
####
|
| 96 |
|
|
|
|
| 97 |
# Create the Gradio interface
|
| 98 |
with gr.Blocks() as interface:
|
| 99 |
gr.Markdown("Emotional Machines test: Load or Record an audio file to speech emotion analysis")
|
|
@@ -105,11 +106,11 @@ with gr.Blocks() as interface:
|
|
| 105 |
submit_button = gr.Button("Submit")
|
| 106 |
output_label = [gr.Label("Prediction"), gr.Image(type='pil')] # Use a single Label instead of a list
|
| 107 |
|
| 108 |
-
|
| 109 |
-
|
| 110 |
|
| 111 |
interface.launch()
|
| 112 |
-
|
| 113 |
|
| 114 |
|
| 115 |
|
|
|
|
| 94 |
return None
|
| 95 |
####
|
| 96 |
|
| 97 |
+
|
| 98 |
# Create the Gradio interface
|
| 99 |
with gr.Blocks() as interface:
|
| 100 |
gr.Markdown("Emotional Machines test: Load or Record an audio file to speech emotion analysis")
|
|
|
|
| 106 |
submit_button = gr.Button("Submit")
|
| 107 |
output_label = [gr.Label("Prediction"), gr.Image(type='pil')] # Use a single Label instead of a list
|
| 108 |
|
| 109 |
+
# Set the function to be called when the button is clicked
|
| 110 |
+
submit_button.click(get_predictions, inputs=input_audio, outputs=output_label, transcribe=transcribe)
|
| 111 |
|
| 112 |
interface.launch()
|
| 113 |
+
|
| 114 |
|
| 115 |
|
| 116 |
|