Muhammad Anas Akhtar
commited on
Update app.py
Browse files
app.py
CHANGED
|
@@ -7,8 +7,10 @@ import scipy.io.wavfile as wavfile
|
|
| 7 |
from transformers import pipeline
|
| 8 |
|
| 9 |
|
|
|
|
|
|
|
| 10 |
narrator = pipeline("text-to-speech",
|
| 11 |
-
model="
|
| 12 |
|
| 13 |
object_detector = pipeline("object-detection",
|
| 14 |
model="facebook/detr-resnet-50")
|
|
@@ -130,7 +132,7 @@ def detect_object(image):
|
|
| 130 |
demo = gr.Interface(fn=detect_object,
|
| 131 |
inputs=[gr.Image(label="Select Image",type="pil")],
|
| 132 |
outputs=[gr.Image(label="Processed Image", type="pil"), gr.Audio(label="Generated Audio")],
|
| 133 |
-
title="Object Detector with Audio",
|
| 134 |
description="THIS APPLICATION WILL BE USED TO HIGHLIGHT OBJECTS AND GIVES AUDIO DESCRIPTION FOR THE PROVIDED INPUT IMAGE.")
|
| 135 |
demo.launch()
|
| 136 |
|
|
|
|
| 7 |
from transformers import pipeline
|
| 8 |
|
| 9 |
|
| 10 |
+
|
| 11 |
+
|
| 12 |
narrator = pipeline("text-to-speech",
|
| 13 |
+
model="kakao-enterprise/vits-ljs")
|
| 14 |
|
| 15 |
object_detector = pipeline("object-detection",
|
| 16 |
model="facebook/detr-resnet-50")
|
|
|
|
| 132 |
demo = gr.Interface(fn=detect_object,
|
| 133 |
inputs=[gr.Image(label="Select Image",type="pil")],
|
| 134 |
outputs=[gr.Image(label="Processed Image", type="pil"), gr.Audio(label="Generated Audio")],
|
| 135 |
+
title="@GenAILearniverse Project 7: Object Detector with Audio",
|
| 136 |
description="THIS APPLICATION WILL BE USED TO HIGHLIGHT OBJECTS AND GIVES AUDIO DESCRIPTION FOR THE PROVIDED INPUT IMAGE.")
|
| 137 |
demo.launch()
|
| 138 |
|