Spaces:
Running
Running
Commit
·
54ab0ba
1
Parent(s):
911f938
fix f-string
Browse files
app.py
CHANGED
|
@@ -101,7 +101,7 @@ def save_leaderboard(data: list):
|
|
| 101 |
path_in_repo=LEADERBOARD_FILENAME,
|
| 102 |
repo_id=LEADERBOARD_DATASET,
|
| 103 |
repo_type="dataset",
|
| 104 |
-
commit_message=f"Update leaderboard - {datetime.
|
| 105 |
)
|
| 106 |
print(f"Leaderboard saved to {LEADERBOARD_DATASET}")
|
| 107 |
|
|
@@ -290,8 +290,8 @@ def run_evaluation(
|
|
| 290 |
model_info = hf_model_info(model_id, token=HF_TOKEN)
|
| 291 |
model_last_modified = model_info.lastModified
|
| 292 |
except Exception as e:
|
| 293 |
-
return f"## Evaluation Failed
|
| 294 |
-
Could not fetch model info for `{model_id}`: {e}"
|
| 295 |
|
| 296 |
leaderboard = load_leaderboard()
|
| 297 |
model_entry = next((e for e in leaderboard if e.get("model_id") == model_id), None)
|
|
@@ -399,13 +399,12 @@ which adds the required metadata to the README.md file.
|
|
| 399 |
update_message = "New entry added to leaderboard!"
|
| 400 |
else:
|
| 401 |
old_rate = model_entry.get("legal_rate", 0)
|
| 402 |
-
model_entry.update(new_entry)
|
| 403 |
save_leaderboard(leaderboard)
|
| 404 |
if result.legal_rate_with_retry > old_rate:
|
| 405 |
update_message = f"Improved! {old_rate*100:.1f}% -> {result.legal_rate_with_retry*100:.1f}%"
|
| 406 |
else:
|
| 407 |
update_message = f"Re-evaluated. Previous: {old_rate*100:.1f}%, This run: {result.legal_rate_with_retry*100:.1f}%"
|
| 408 |
-
update_message = f"No improvement. Best: {old_rate*100:.1f}%, This run: {result.legal_rate*100:.1f}%"
|
| 409 |
|
| 410 |
# Post discussion to model page
|
| 411 |
if HF_TOKEN:
|
|
|
|
| 101 |
path_in_repo=LEADERBOARD_FILENAME,
|
| 102 |
repo_id=LEADERBOARD_DATASET,
|
| 103 |
repo_type="dataset",
|
| 104 |
+
commit_message=f"Update leaderboard - {datetime.utcnow().strftime('%Y-%m-%d %H:%M')}",
|
| 105 |
)
|
| 106 |
print(f"Leaderboard saved to {LEADERBOARD_DATASET}")
|
| 107 |
|
|
|
|
| 290 |
model_info = hf_model_info(model_id, token=HF_TOKEN)
|
| 291 |
model_last_modified = model_info.lastModified
|
| 292 |
except Exception as e:
|
| 293 |
+
return f"""## Evaluation Failed
|
| 294 |
+
Could not fetch model info for `{model_id}`: {e}"""
|
| 295 |
|
| 296 |
leaderboard = load_leaderboard()
|
| 297 |
model_entry = next((e for e in leaderboard if e.get("model_id") == model_id), None)
|
|
|
|
| 399 |
update_message = "New entry added to leaderboard!"
|
| 400 |
else:
|
| 401 |
old_rate = model_entry.get("legal_rate", 0)
|
| 402 |
+
model_entry.update(new_entry) # Update existing entry for the model
|
| 403 |
save_leaderboard(leaderboard)
|
| 404 |
if result.legal_rate_with_retry > old_rate:
|
| 405 |
update_message = f"Improved! {old_rate*100:.1f}% -> {result.legal_rate_with_retry*100:.1f}%"
|
| 406 |
else:
|
| 407 |
update_message = f"Re-evaluated. Previous: {old_rate*100:.1f}%, This run: {result.legal_rate_with_retry*100:.1f}%"
|
|
|
|
| 408 |
|
| 409 |
# Post discussion to model page
|
| 410 |
if HF_TOKEN:
|