++
Browse files- app.py +3 -5
- src/envs.py +0 -1
app.py
CHANGED
|
@@ -49,7 +49,7 @@ def init_leaderboard(dataframe):
|
|
| 49 |
cant_deselect=[c.name for c in fields(AutoEvalColumn) if c.never_hidden],
|
| 50 |
label="Select Columns to Display:",
|
| 51 |
),
|
| 52 |
-
search_columns=[AutoEvalColumn.model_name.name],
|
| 53 |
hide_columns=[c.name for c in fields(AutoEvalColumn) if c.hidden],
|
| 54 |
filter_columns=[],
|
| 55 |
bool_checkboxgroup_label="Hide models",
|
|
@@ -64,15 +64,12 @@ with demo:
|
|
| 64 |
|
| 65 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
| 66 |
with gr.TabItem("🏅 GridNet-HD Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
| 67 |
-
# leaderboard_container = gr.Column()
|
| 68 |
leaderboard = init_leaderboard(LEADERBOARD_DF)
|
| 69 |
-
|
| 70 |
-
|
| 71 |
def reload_leaderboard():
|
| 72 |
# Reload dataframe
|
| 73 |
print("reload_leaderboard")
|
| 74 |
df = get_leaderboard_df(EVAL_RESULTS_PATH, COLS)
|
| 75 |
-
#print("ok")
|
| 76 |
return df
|
| 77 |
# Load on app start or page refresh
|
| 78 |
demo.load(
|
|
@@ -80,6 +77,7 @@ with demo:
|
|
| 80 |
inputs=[],
|
| 81 |
outputs=[leaderboard]
|
| 82 |
)
|
|
|
|
| 83 |
|
| 84 |
with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
|
| 85 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
|
|
|
| 49 |
cant_deselect=[c.name for c in fields(AutoEvalColumn) if c.never_hidden],
|
| 50 |
label="Select Columns to Display:",
|
| 51 |
),
|
| 52 |
+
search_columns=[AutoEvalColumn.model_name.name,AutoEvalColumn.eval_name.name],
|
| 53 |
hide_columns=[c.name for c in fields(AutoEvalColumn) if c.hidden],
|
| 54 |
filter_columns=[],
|
| 55 |
bool_checkboxgroup_label="Hide models",
|
|
|
|
| 64 |
|
| 65 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
| 66 |
with gr.TabItem("🏅 GridNet-HD Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
|
|
|
| 67 |
leaderboard = init_leaderboard(LEADERBOARD_DF)
|
| 68 |
+
"""
|
|
|
|
| 69 |
def reload_leaderboard():
|
| 70 |
# Reload dataframe
|
| 71 |
print("reload_leaderboard")
|
| 72 |
df = get_leaderboard_df(EVAL_RESULTS_PATH, COLS)
|
|
|
|
| 73 |
return df
|
| 74 |
# Load on app start or page refresh
|
| 75 |
demo.load(
|
|
|
|
| 77 |
inputs=[],
|
| 78 |
outputs=[leaderboard]
|
| 79 |
)
|
| 80 |
+
"""
|
| 81 |
|
| 82 |
with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
|
| 83 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
src/envs.py
CHANGED
|
@@ -10,7 +10,6 @@ OWNER = "heig-vd-geo" # Change to your org - don't forget to create a results an
|
|
| 10 |
# ----------------------------------
|
| 11 |
|
| 12 |
REPO_ID = f"{OWNER}/GridNet-HD-Leaderboard"
|
| 13 |
-
# QUEUE_REPO = f"{OWNER}/requests"
|
| 14 |
RESULTS_REPO = f"{OWNER}/GridNet-HD-results"
|
| 15 |
|
| 16 |
# If you setup a cache later, just change HF_HOME
|
|
|
|
| 10 |
# ----------------------------------
|
| 11 |
|
| 12 |
REPO_ID = f"{OWNER}/GridNet-HD-Leaderboard"
|
|
|
|
| 13 |
RESULTS_REPO = f"{OWNER}/GridNet-HD-results"
|
| 14 |
|
| 15 |
# If you setup a cache later, just change HF_HOME
|