Spaces:
Running
on
Zero
Running
on
Zero
fix demo
Browse files
app.py
CHANGED
|
@@ -232,7 +232,6 @@ with gr.Blocks(theme=theme, title="HORT: Monocular Hand-held Objects Reconstruct
|
|
| 232 |
with gr.Row():
|
| 233 |
with gr.Column():
|
| 234 |
input_image = gr.Image(label="Input image", type="numpy")
|
| 235 |
-
threshold = gr.Slider(value=0.3, minimum=0.05, maximum=0.95, step=0.05, label='Detection Confidence Threshold')
|
| 236 |
submit = gr.Button("Submit", variant="primary")
|
| 237 |
example_images = gr.Examples([
|
| 238 |
['/home/user/app/assets/test1.png'],
|
|
@@ -248,9 +247,9 @@ with gr.Blocks(theme=theme, title="HORT: Monocular Hand-held Objects Reconstruct
|
|
| 248 |
|
| 249 |
with gr.Column():
|
| 250 |
reconstruction = gr.Image(label="Reconstructions", type="numpy")
|
| 251 |
-
output_meshes = gr.Model3D(height=300, zoom_speed=0.5, pan_speed=0.5)
|
| 252 |
hands_detected = gr.Textbox(label="Hands Detected")
|
| 253 |
|
| 254 |
-
submit.click(fn=render_reconstruction, inputs=[input_image
|
| 255 |
|
| 256 |
demo.launch(share=True)
|
|
|
|
| 232 |
with gr.Row():
|
| 233 |
with gr.Column():
|
| 234 |
input_image = gr.Image(label="Input image", type="numpy")
|
|
|
|
| 235 |
submit = gr.Button("Submit", variant="primary")
|
| 236 |
example_images = gr.Examples([
|
| 237 |
['/home/user/app/assets/test1.png'],
|
|
|
|
| 247 |
|
| 248 |
with gr.Column():
|
| 249 |
reconstruction = gr.Image(label="Reconstructions", type="numpy")
|
| 250 |
+
output_meshes = gr.Model3D(label="3D Models", height=300, zoom_speed=0.5, pan_speed=0.5)
|
| 251 |
hands_detected = gr.Textbox(label="Hands Detected")
|
| 252 |
|
| 253 |
+
submit.click(fn=render_reconstruction, inputs=[input_image], outputs=[reconstruction, hands_detected, output_meshes])
|
| 254 |
|
| 255 |
demo.launch(share=True)
|