Upload app.py with huggingface_hub
Browse files
app.py
CHANGED
|
@@ -40,7 +40,6 @@ print("📄 Files:", [p.name for p in LOCAL_DATA_DIR.iterdir()])
|
|
| 40 |
OMNI_MLLM_RENAME = {
|
| 41 |
"Task1-1": "Image\nUnderstanding",
|
| 42 |
"Task1-2": "Video\nUnderstanding",
|
| 43 |
-
"Task1-3": "Audio\nReasoning",
|
| 44 |
|
| 45 |
"Task2-1": "Speech-Driven\nImage Understanding",
|
| 46 |
"Task2-2": "Image-Audio\nReasoning",
|
|
@@ -56,6 +55,10 @@ OMNI_MLLM_RENAME = {
|
|
| 56 |
"Task3-6": "Next-Action\nPrediction",
|
| 57 |
}
|
| 58 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
IMAGE_GEN_RENAME = {
|
| 60 |
"WIScore": "WIScore",
|
| 61 |
"SC": "Semantic\nConsistency",
|
|
@@ -290,6 +293,19 @@ with gr.Blocks(
|
|
| 290 |
value=df_vid,
|
| 291 |
interactive=False,
|
| 292 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 293 |
|
| 294 |
# ---------- Refresh ----------
|
| 295 |
gr.Button("🔄 Refresh All").click(
|
|
@@ -297,8 +313,9 @@ with gr.Blocks(
|
|
| 297 |
load_csv("omni-mllm.csv", "Overall").rename(columns=OMNI_MLLM_RENAME),
|
| 298 |
load_csv("image-gen.csv", "Overall").rename(columns=IMAGE_GEN_RENAME),
|
| 299 |
load_csv("video-gen.csv", "Overall").rename(columns=VIDEO_GEN_RENAME),
|
|
|
|
| 300 |
),
|
| 301 |
-
outputs=[omni_table, image_table, video_table],
|
| 302 |
)
|
| 303 |
|
| 304 |
demo.launch()
|
|
|
|
| 40 |
OMNI_MLLM_RENAME = {
|
| 41 |
"Task1-1": "Image\nUnderstanding",
|
| 42 |
"Task1-2": "Video\nUnderstanding",
|
|
|
|
| 43 |
|
| 44 |
"Task2-1": "Speech-Driven\nImage Understanding",
|
| 45 |
"Task2-2": "Image-Audio\nReasoning",
|
|
|
|
| 55 |
"Task3-6": "Next-Action\nPrediction",
|
| 56 |
}
|
| 57 |
|
| 58 |
+
AUDIO_RENAME = {
|
| 59 |
+
"Task1-3": "Audio\nReasoning"
|
| 60 |
+
}
|
| 61 |
+
|
| 62 |
IMAGE_GEN_RENAME = {
|
| 63 |
"WIScore": "WIScore",
|
| 64 |
"SC": "Semantic\nConsistency",
|
|
|
|
| 293 |
value=df_vid,
|
| 294 |
interactive=False,
|
| 295 |
)
|
| 296 |
+
|
| 297 |
+
# ---------- Audio Reasoning ----------
|
| 298 |
+
with gr.Tab("🎵 Audio Reasoning"):
|
| 299 |
+
gr.Markdown("Evaluation results for audio reasoning models.")
|
| 300 |
+
|
| 301 |
+
df_aud = load_csv("audio-reasoning.csv", sort_key="Task1-3")
|
| 302 |
+
df_aud = df_aud.rename(columns=AUDIO_RENAME)
|
| 303 |
+
|
| 304 |
+
audio_table = gr.Dataframe(
|
| 305 |
+
value=df_aud,
|
| 306 |
+
interactive=False,
|
| 307 |
+
)
|
| 308 |
+
|
| 309 |
|
| 310 |
# ---------- Refresh ----------
|
| 311 |
gr.Button("🔄 Refresh All").click(
|
|
|
|
| 313 |
load_csv("omni-mllm.csv", "Overall").rename(columns=OMNI_MLLM_RENAME),
|
| 314 |
load_csv("image-gen.csv", "Overall").rename(columns=IMAGE_GEN_RENAME),
|
| 315 |
load_csv("video-gen.csv", "Overall").rename(columns=VIDEO_GEN_RENAME),
|
| 316 |
+
load_csv("audio-reasoning.csv", "Task1-3").rename(columns=AUDIO_RENAME),
|
| 317 |
),
|
| 318 |
+
outputs=[omni_table, image_table, video_table, audio_table],
|
| 319 |
)
|
| 320 |
|
| 321 |
demo.launch()
|