|
|
import argparse |
|
|
import json |
|
|
import os |
|
|
import sys |
|
|
import traceback |
|
|
import importlib.util |
|
|
import pandas as pd |
|
|
import gc |
|
|
import time |
|
|
|
|
|
|
|
|
class Evaluator: |
|
|
def __init__(self, problem_dir: str): |
|
|
self.problem_dir = problem_dir |
|
|
self.resources_dir = os.path.join(problem_dir, "resources") |
|
|
|
|
|
mounted_datasets_dir = "/datasets/llm_sql/large" |
|
|
if os.path.exists(mounted_datasets_dir) and os.listdir(mounted_datasets_dir): |
|
|
self.datasets_dir = mounted_datasets_dir |
|
|
else: |
|
|
|
|
|
self.datasets_dir = os.path.join(self.resources_dir, "datasets") |
|
|
ordered_names = ["PDMX.csv", "credit.csv"] |
|
|
self.trace_files = [ |
|
|
os.path.join(self.datasets_dir, name) |
|
|
for name in ordered_names |
|
|
if os.path.exists(os.path.join(self.datasets_dir, name)) |
|
|
] |
|
|
|
|
|
|
|
|
self.baseline_cache_file = os.path.join(self.problem_dir, "baseline_cache.json") |
|
|
|
|
|
|
|
|
self.col_merges = [ |
|
|
[["path", "metadata"], ["hasmetadata", "isofficial", "isuserpublisher", "isdraft", "hasannotations", "subsetall"]], |
|
|
[ |
|
|
["PAY_0", "PAY_2", "PAY_3", "PAY_4", "PAY_5", "PAY_6"], |
|
|
["BILL_AMT1", "BILL_AMT2", "BILL_AMT3", "BILL_AMT4", "BILL_AMT5", "BILL_AMT6"], |
|
|
["PAY_AMT1", "PAY_AMT2", "PAY_AMT3", "PAY_AMT4", "PAY_AMT5", "PAY_AMT6"], |
|
|
["SEX", "EDUCATION", "MARRIAGE", "AGE"], |
|
|
], |
|
|
] |
|
|
|
|
|
|
|
|
if self.resources_dir not in sys.path: |
|
|
sys.path.insert(0, self.resources_dir) |
|
|
|
|
|
from utils import evaluate_df_prefix_hit_cnt |
|
|
self._eval_prefix = evaluate_df_prefix_hit_cnt |
|
|
|
|
|
def _calculate_baseline_hit_rate(self) -> float: |
|
|
"""Calculate the baseline hit rate using original column order (0-point anchor)""" |
|
|
|
|
|
baseline_hit_rates = [] |
|
|
for csv_path, merge_spec in zip(self.trace_files, self.col_merges[: len(self.trace_files)]): |
|
|
dataset_name = os.path.basename(csv_path) |
|
|
|
|
|
|
|
|
df = pd.read_csv(csv_path, low_memory=False) |
|
|
|
|
|
|
|
|
if merge_spec: |
|
|
for col_to_merge in merge_spec: |
|
|
if all(col in df.columns for col in col_to_merge): |
|
|
merged_name = "_".join(col_to_merge) |
|
|
df[merged_name] = df[col_to_merge].apply( |
|
|
lambda x: "".join([f"{val}" for val in x]), axis=1 |
|
|
) |
|
|
df = df.drop(columns=col_to_merge) |
|
|
|
|
|
|
|
|
_, hit_rate_percent = self._eval_prefix(df) |
|
|
hit_rate = hit_rate_percent / 100.0 |
|
|
baseline_hit_rates.append(hit_rate) |
|
|
|
|
|
|
|
|
del df |
|
|
gc.collect() |
|
|
|
|
|
avg_baseline_hit = sum(baseline_hit_rates) / len(self.trace_files) if baseline_hit_rates else 0.0 |
|
|
|
|
|
|
|
|
try: |
|
|
cache_data = { |
|
|
'baseline_hit_rate': avg_baseline_hit, |
|
|
'individual_rates': baseline_hit_rates, |
|
|
'num_datasets': len(self.trace_files), |
|
|
} |
|
|
os.makedirs(os.path.dirname(self.baseline_cache_file) or ".", exist_ok=True) |
|
|
with open(self.baseline_cache_file, 'w') as f: |
|
|
json.dump(cache_data, f, indent=2) |
|
|
except Exception as e: |
|
|
print(f"[WARNING] Failed to save baseline cache: {e}") |
|
|
|
|
|
return avg_baseline_hit |
|
|
|
|
|
def evaluate(self, solution_module_path: str) -> dict: |
|
|
spec = importlib.util.spec_from_file_location("solution", solution_module_path) |
|
|
solution = importlib.util.module_from_spec(spec) |
|
|
spec.loader.exec_module(solution) |
|
|
|
|
|
if not hasattr(solution, "Solution"): |
|
|
return {"score": 0.0, "runs_successfully": 0.0, "error": "Missing Solution"} |
|
|
|
|
|
solver = solution.Solution() |
|
|
|
|
|
|
|
|
baseline_hit = None |
|
|
if os.path.exists(self.baseline_cache_file): |
|
|
try: |
|
|
with open(self.baseline_cache_file, 'r') as f: |
|
|
cache_data = json.load(f) |
|
|
baseline_hit = cache_data.get('baseline_hit_rate') |
|
|
except Exception: |
|
|
baseline_hit = None |
|
|
if baseline_hit is None: |
|
|
baseline_hit = self._calculate_baseline_hit_rate() |
|
|
|
|
|
hit_rates = [] |
|
|
total_runtime = 0.0 |
|
|
for csv_path, merge_spec in zip(self.trace_files, self.col_merges[: len(self.trace_files)]): |
|
|
dataset_name = os.path.basename(csv_path) |
|
|
|
|
|
df = pd.read_csv(csv_path, low_memory=False) |
|
|
|
|
|
start = time.time() |
|
|
reordered = solver.solve( |
|
|
df, |
|
|
early_stop=100000, |
|
|
row_stop=4, |
|
|
col_stop=2, |
|
|
col_merge=merge_spec, |
|
|
one_way_dep=[], |
|
|
distinct_value_threshold=0.7, |
|
|
parallel=True, |
|
|
) |
|
|
runtime = time.time() - start |
|
|
total_runtime += runtime |
|
|
_, hit_rate_percent = self._eval_prefix(reordered) |
|
|
hit_rates.append(hit_rate_percent / 100.0) |
|
|
|
|
|
|
|
|
del df |
|
|
del reordered |
|
|
gc.collect() |
|
|
|
|
|
if not self.trace_files: |
|
|
return {"score": 0.0, "runs_successfully": 0.0, "error": "No datasets found"} |
|
|
|
|
|
avg_runtime = total_runtime / len(self.trace_files) |
|
|
|
|
|
|
|
|
if avg_runtime > 10.0: |
|
|
score = 0.0 |
|
|
else: |
|
|
individual_scores = [] |
|
|
for i, hit_rate in enumerate(hit_rates): |
|
|
dataset_score = ((hit_rate - baseline_hit) / (1.0 - baseline_hit)) * 100 |
|
|
dataset_score = max(0, min(100, dataset_score)) |
|
|
individual_scores.append(dataset_score) |
|
|
score = sum(individual_scores) / len(individual_scores) |
|
|
avg_hit = sum(hit_rates) / len(self.trace_files) |
|
|
|
|
|
return {"score": score, "runs_successfully": 1.0, "avg_hit_rate": sum(hit_rates) / len(self.trace_files) * 100 if hit_rates else 0.0, "total_runtime": total_runtime, "avg_runtime": avg_runtime, "runtime_threshold": 10.0} |
|
|
|
|
|
|
|
|
def main(): |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument("--solution", required=True) |
|
|
parser.add_argument("--out", required=True) |
|
|
args = parser.parse_args() |
|
|
|
|
|
try: |
|
|
problem_dir = os.path.dirname(os.path.abspath(__file__)) |
|
|
result = Evaluator(problem_dir).evaluate(args.solution) |
|
|
except Exception as e: |
|
|
print(f"[evaluator] ERROR: {e}", file=sys.stderr) |
|
|
print(traceback.format_exc(), file=sys.stderr) |
|
|
result = {"score": 0.0, "runs_successfully": 0.0, "error": str(e)} |
|
|
|
|
|
os.makedirs(os.path.dirname(args.out) or ".", exist_ok=True) |
|
|
with open(args.out, "w") as f: |
|
|
json.dump(result, f) |
|
|
print(json.dumps(result)) |
|
|
return 0 |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
raise SystemExit(main()) |
|
|
|
|
|
|
|
|
|