Spaces:
Running
Running
ycy
commited on
Commit
·
f8c4b7a
1
Parent(s):
aed219d
test
Browse files- src/display/utils.py +1 -1
- src/leaderboard/read_evals.py +5 -4
- src/populate.py +2 -1
src/display/utils.py
CHANGED
|
@@ -36,7 +36,7 @@ auto_eval_column_dict.append(["architecture", ColumnContent, ColumnContent("Arch
|
|
| 36 |
auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
|
| 37 |
auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
|
| 38 |
auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
|
| 39 |
-
auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "
|
| 40 |
auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
|
| 41 |
auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
|
| 42 |
auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
|
|
|
|
| 36 |
auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
|
| 37 |
auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
|
| 38 |
auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
|
| 39 |
+
auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "float", False)])
|
| 40 |
auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
|
| 41 |
auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
|
| 42 |
auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
|
src/leaderboard/read_evals.py
CHANGED
|
@@ -40,13 +40,14 @@ class EvalResult:
|
|
| 40 |
|
| 41 |
config = data.get("config")
|
| 42 |
|
| 43 |
-
# Precision
|
| 44 |
precision = Precision.from_str(config.get("model_dtype"))
|
| 45 |
|
| 46 |
-
# Get model and org
|
| 47 |
org_and_model = config.get("model_name", config.get("model_args", None))
|
| 48 |
org_and_model = org_and_model.split("/", 1)
|
| 49 |
|
|
|
|
| 50 |
if len(org_and_model) == 1:
|
| 51 |
org = None
|
| 52 |
model = org_and_model[0]
|
|
@@ -114,8 +115,8 @@ class EvalResult:
|
|
| 114 |
#average = next(iter(self.results.values()))
|
| 115 |
data_dict = {
|
| 116 |
"eval_name": self.eval_name, # not a column, just a save name,
|
| 117 |
-
AutoEvalColumn.precision.name: self.precision.value.name,
|
| 118 |
-
AutoEvalColumn.model_type.name: self.model_type.value.name,
|
| 119 |
AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
|
| 120 |
AutoEvalColumn.weight_type.name: self.weight_type.value.name,
|
| 121 |
AutoEvalColumn.architecture.name: self.architecture,
|
|
|
|
| 40 |
|
| 41 |
config = data.get("config")
|
| 42 |
|
| 43 |
+
# Precision #1
|
| 44 |
precision = Precision.from_str(config.get("model_dtype"))
|
| 45 |
|
| 46 |
+
# Get model and org
|
| 47 |
org_and_model = config.get("model_name", config.get("model_args", None))
|
| 48 |
org_and_model = org_and_model.split("/", 1)
|
| 49 |
|
| 50 |
+
# 如果model_name = “”
|
| 51 |
if len(org_and_model) == 1:
|
| 52 |
org = None
|
| 53 |
model = org_and_model[0]
|
|
|
|
| 115 |
#average = next(iter(self.results.values()))
|
| 116 |
data_dict = {
|
| 117 |
"eval_name": self.eval_name, # not a column, just a save name,
|
| 118 |
+
AutoEvalColumn.precision.name: self.precision.value.name, #results/precision
|
| 119 |
+
AutoEvalColumn.model_type.name: self.model_type.value.name, #requests/model_type
|
| 120 |
AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
|
| 121 |
AutoEvalColumn.weight_type.name: self.weight_type.value.name,
|
| 122 |
AutoEvalColumn.architecture.name: self.architecture,
|
src/populate.py
CHANGED
|
@@ -13,7 +13,8 @@ def get_leaderboard_df(results_path: str, requests_path: str, cols: list, benchm
|
|
| 13 |
raw_data = get_raw_eval_results(results_path, requests_path)
|
| 14 |
|
| 15 |
all_data_json = [v.to_dict() for v in raw_data]
|
| 16 |
-
|
|
|
|
| 17 |
df = pd.DataFrame.from_records(all_data_json)
|
| 18 |
|
| 19 |
df = df.sort_values(by=[AutoEvalColumn.task0.name], ascending=False)
|
|
|
|
| 13 |
raw_data = get_raw_eval_results(results_path, requests_path)
|
| 14 |
|
| 15 |
all_data_json = [v.to_dict() for v in raw_data]
|
| 16 |
+
print(all_data_json)
|
| 17 |
+
assert 0
|
| 18 |
df = pd.DataFrame.from_records(all_data_json)
|
| 19 |
|
| 20 |
df = df.sort_values(by=[AutoEvalColumn.task0.name], ascending=False)
|