Spaces:
Running
Running
update
Browse files
app.py
CHANGED
|
@@ -12,27 +12,27 @@ from src.about import (
|
|
| 12 |
|
| 13 |
data = {
|
| 14 |
"Method": [
|
| 15 |
-
|
| 16 |
-
|
| 17 |
-
|
| 18 |
-
"RAG (E5
|
| 19 |
"RAG (E5) + LM Rerank (Llama-3.1-70B)",
|
| 20 |
"Handwritten LOTUS (GPT-4o)",
|
| 21 |
"Zero-shot Text2SQL (GPT-4o)",
|
| 22 |
-
|
| 23 |
-
"RAG (GPT-4o)",
|
| 24 |
"RAG (E5) + LM Rerank (GPT-4o)",
|
| 25 |
"Handwritten LOTUS (o3-mini)",
|
| 26 |
"Zero-shot Text2SQL (o3-mini)",
|
| 27 |
-
|
| 28 |
-
"RAG (o3-mini)",
|
| 29 |
"RAG (E5) + LM Rerank (o3-mini)",
|
| 30 |
"Zero-shot Text2SQL (Deepseek-R1)",
|
| 31 |
-
|
| 32 |
],
|
| 33 |
# "Model": ["meta-llama/Llama-3.1-70B"] * 5,
|
| 34 |
-
|
| 35 |
-
"Execution Accuracy": [0.0, 2.0, 55.0, 18.0, 3.0, 3.0, 65.0, 18.0, 7.0, 7.0, 12.0],
|
| 36 |
}
|
| 37 |
|
| 38 |
leaderboard_df = pd.DataFrame(data)
|
|
@@ -43,7 +43,7 @@ leaderboard_df = leaderboard_df.sort_values(
|
|
| 43 |
leaderboard_df.insert(0, "Rank", leaderboard_df.index - 1)
|
| 44 |
leaderboard_df.loc[0, "Rank"] = ""
|
| 45 |
leaderboard_df.loc[1, "Rank"] = ""
|
| 46 |
-
|
| 47 |
|
| 48 |
def hyperlink_method(method):
|
| 49 |
base_url = "https://github.com/TAG-Research/TAG-Bench/tree/main"
|
|
|
|
| 12 |
|
| 13 |
data = {
|
| 14 |
"Method": [
|
| 15 |
+
"Handwritten LOTUS (Llama-3.1-70B)",
|
| 16 |
+
"Zero-shot Text2SQL (Llama-3.1-70B)",
|
| 17 |
+
"Zero-shot Text2SQL + LM Generation (Llama-3.1-70B)",
|
| 18 |
+
"RAG (E5 + Llama-3.1-70B)",
|
| 19 |
"RAG (E5) + LM Rerank (Llama-3.1-70B)",
|
| 20 |
"Handwritten LOTUS (GPT-4o)",
|
| 21 |
"Zero-shot Text2SQL (GPT-4o)",
|
| 22 |
+
"Zero-shot Text2SQL + LM Generation (GPT-4o)",
|
| 23 |
+
"RAG (E5 + GPT-4o)",
|
| 24 |
"RAG (E5) + LM Rerank (GPT-4o)",
|
| 25 |
"Handwritten LOTUS (o3-mini)",
|
| 26 |
"Zero-shot Text2SQL (o3-mini)",
|
| 27 |
+
"Zero-shot Text2SQL + LM Generation (o3-mini)",
|
| 28 |
+
"RAG (E5 + o3-mini)",
|
| 29 |
"RAG (E5) + LM Rerank (o3-mini)",
|
| 30 |
"Zero-shot Text2SQL (Deepseek-R1)",
|
| 31 |
+
"Zero-shot Text2SQL + LM Generation (Deepseek-R1)",
|
| 32 |
],
|
| 33 |
# "Model": ["meta-llama/Llama-3.1-70B"] * 5,
|
| 34 |
+
"Execution Accuracy": [55.0, 17.0, 13.0, 0.0, 2.0, 55.0, 18.0, 15.0, 3.0, 3.0, 65.0, 18.0, 30.0, 7.0, 7.0, 12.0, 0.0],
|
| 35 |
+
# "Execution Accuracy": [0.0, 2.0, 55.0, 18.0, 3.0, 3.0, 65.0, 18.0, 7.0, 7.0, 12.0],
|
| 36 |
}
|
| 37 |
|
| 38 |
leaderboard_df = pd.DataFrame(data)
|
|
|
|
| 43 |
leaderboard_df.insert(0, "Rank", leaderboard_df.index - 1)
|
| 44 |
leaderboard_df.loc[0, "Rank"] = ""
|
| 45 |
leaderboard_df.loc[1, "Rank"] = ""
|
| 46 |
+
leaderboard_df.loc[2, "Rank"] = ""
|
| 47 |
|
| 48 |
def hyperlink_method(method):
|
| 49 |
base_url = "https://github.com/TAG-Research/TAG-Bench/tree/main"
|