Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -205,13 +205,9 @@ with gr.Blocks(css=css, theme="bethecloud/storj_theme") as demo:
|
|
| 205 |
with gr.Column():
|
| 206 |
with gr.Column(elem_classes="canvas-output"):
|
| 207 |
gr.Markdown("## Output")
|
| 208 |
-
output = gr.Textbox(label="Raw Output Stream", interactive=False, lines=
|
| 209 |
with gr.Accordion("(Result.md)", open=False):
|
| 210 |
markdown_output = gr.Markdown(label="(Result.Md)")
|
| 211 |
-
gr.Markdown("**Model Info 💻** | [Report Bug](https://huggingface.co/spaces/prithivMLmods/Multimodal-VLM-Thinking/discussions)")
|
| 212 |
-
gr.Markdown("> Using **[Qwen/Qwen3-VL-30B-A3B-Instruct](https://huggingface.co/Qwen/Qwen3-VL-30B-A3B-Instruct)**, a powerful and versatile vision-language model. It excels at understanding and processing both text and visual information, making it suitable for a wide range of multimodal tasks. The model demonstrates strong performance in areas like visual question answering, image captioning, and video analysis.")
|
| 213 |
-
gr.Markdown("> ⚠️ Note: Video inference performance can vary depending on the complexity and length of the video.")
|
| 214 |
-
|
| 215 |
image_submit.click(
|
| 216 |
fn=generate_image,
|
| 217 |
inputs=[image_query, image_upload, max_new_tokens, temperature, top_p, top_k, repetition_penalty],
|
|
|
|
| 205 |
with gr.Column():
|
| 206 |
with gr.Column(elem_classes="canvas-output"):
|
| 207 |
gr.Markdown("## Output")
|
| 208 |
+
output = gr.Textbox(label="Raw Output Stream", interactive=False, lines=9, show_copy_button=True)
|
| 209 |
with gr.Accordion("(Result.md)", open=False):
|
| 210 |
markdown_output = gr.Markdown(label="(Result.Md)")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 211 |
image_submit.click(
|
| 212 |
fn=generate_image,
|
| 213 |
inputs=[image_query, image_upload, max_new_tokens, temperature, top_p, top_k, repetition_penalty],
|