Spaces:
Runtime error
Runtime error
Update run_eval.py
Browse files- run_eval.py +8 -6
run_eval.py
CHANGED
|
@@ -44,7 +44,7 @@ for cfg in CONFIGS:
|
|
| 44 |
adapter_type = cfg.get("adapter_type", "LoRA")
|
| 45 |
tasks = cfg["tasks"]
|
| 46 |
|
| 47 |
-
print(f"\
|
| 48 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id, use_fast=True)
|
| 49 |
|
| 50 |
# Try causal first, fallback to encoder
|
|
@@ -110,14 +110,16 @@ with tempfile.TemporaryDirectory() as tmp:
|
|
| 110 |
df_existing = pd.read_parquet(current_path)
|
| 111 |
df_combined = pd.concat([df_existing, df_new], ignore_index=True)
|
| 112 |
|
| 113 |
-
df_combined = (
|
| 114 |
-
|
| 115 |
-
.sort_values("run_date")
|
| 116 |
-
.drop_duplicates(subset=["model_id", "task", "metric"], keep="last")
|
| 117 |
-
)
|
| 118 |
|
| 119 |
df_combined["value"] = pd.to_numeric(df_combined["value"], errors="coerce")
|
| 120 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 121 |
out = Path("peft_bench.parquet")
|
| 122 |
df_combined.to_parquet(out, index=False)
|
| 123 |
|
|
|
|
| 44 |
adapter_type = cfg.get("adapter_type", "LoRA")
|
| 45 |
tasks = cfg["tasks"]
|
| 46 |
|
| 47 |
+
print(f"\nLoading base model: {base_model_id}")
|
| 48 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id, use_fast=True)
|
| 49 |
|
| 50 |
# Try causal first, fallback to encoder
|
|
|
|
| 110 |
df_existing = pd.read_parquet(current_path)
|
| 111 |
df_combined = pd.concat([df_existing, df_new], ignore_index=True)
|
| 112 |
|
| 113 |
+
df_combined = df_combined.sort_values("run_date")
|
| 114 |
+
|
|
|
|
|
|
|
|
|
|
| 115 |
|
| 116 |
df_combined["value"] = pd.to_numeric(df_combined["value"], errors="coerce")
|
| 117 |
|
| 118 |
+
print("Existing rows:", len(df_existing))
|
| 119 |
+
print("New rows:", len(df_new))
|
| 120 |
+
print("Combined rows (pre-dedup):", len(df_existing) + len(df_new))
|
| 121 |
+
print("Final rows (after dedup):", len(df_combined))
|
| 122 |
+
|
| 123 |
out = Path("peft_bench.parquet")
|
| 124 |
df_combined.to_parquet(out, index=False)
|
| 125 |
|