| import re |
|
|
| import itertools |
| import json |
| import os |
| from concurrent.futures import ThreadPoolExecutor |
| from typing import get_args, Literal, List |
|
|
| import pandas as pd |
| import time |
| from collections import defaultdict, Counter |
| from datetime import datetime |
|
|
| import gradio as gr |
| from huggingface_hub import cached_assets_path |
|
|
| from viewer.agg_score_metrics import agg_score_metrics_dict_prob, custom_task_aggregate_groups, agg_score_metrics |
| from viewer.literals import BASLINE_RUN_NAME, CHECK_MISSING_DATAPOINTS_BUTTON_CLOSE_LABEL, \ |
| CHECK_MISSING_DATAPOINTS_BUTTON_LABEL, \ |
| FALLBACK_TOKEN_NAME, REFERENCE_RUNS |
| from viewer.utils import BASELINE_GROUPING_MODE, create_df_from_run_data, get_run_name_seed, RunInfo, TaskInfo, get_groupped_score, RunData, is_aggregate_column, is_baseline_run, is_reference_run, is_task_column, rescale_scores, select_runs, z_score_normalize |
| from datatrove.io import DataFolder |
| from viewer.task_type_mapping import get_task_type, TASK_TYPE |
| import tqdm as progress |
|
|
| NormalizationMode = Literal["No adjustment", "Rescale", "Z-norm"] |
|
|
| def fetch_run_results(results_uri, runs_to_fetch, steps_to_fetch, |
| oauth_token: gr.OAuthToken | None = None): |
| token = os.environ.get(FALLBACK_TOKEN_NAME) |
| if oauth_token: |
| token = oauth_token.token |
| if not runs_to_fetch: |
| return None, None |
|
|
| steps_to_fetch_list = steps_to_fetch.split(",") |
| data_folder = DataFolder(results_uri, token=token) |
|
|
| def fetch_run_files(run_to_fetch): |
| def filename_to_steps_timestamp(fn): |
| step, ts = fn.split("/results_") |
| dt = datetime.strptime(ts.split(".")[0], "%Y-%m-%dT%H-%M-%S") |
| return int(step), dt |
|
|
| run_path = f"results/{run_to_fetch}" |
|
|
| try: |
| eval_files = [f for f in data_folder.list_files(run_path, recursive=True)] |
| except FileNotFoundError: |
| return [] |
|
|
| |
| step_files = defaultdict(list) |
| for fn in eval_files: |
| steps, ts = filename_to_steps_timestamp(os.path.relpath(fn, run_path)) |
| step_files[steps].append((ts, fn)) |
|
|
| |
| for step in step_files: |
| step_files[step].sort(reverse=True) |
|
|
| |
| results = [] |
| for step, files in step_files.items(): |
| if any(step_element_match(step, step_el) for step_el in steps_to_fetch_list): |
| results.append((run_to_fetch, step, files)) |
| return results |
|
|
| def get_file_with_retry(data_folder: DataFolder, filename: str): |
| save_path = os.path.join(cached_assets_path(library_name="results-viewer", |
| namespace=data_folder.path), filename) |
| if os.path.exists(save_path): |
| with open(save_path, "rb") as f: |
| return f.read() |
|
|
| wait = 1.5 |
| max_retries = 20 |
| for attempt in range(max_retries): |
| try: |
| with data_folder.open(filename, "rb") as f: |
| data = f.read() |
| os.makedirs(os.path.dirname(save_path), exist_ok=True) |
| with open(save_path, "wb") as f: |
| f.write(data) |
| return data |
| except Exception as e: |
| print(f"Error downloading (attempt {attempt + 1}/{max_retries}): {e}") |
| if attempt == max_retries - 1: |
| raise e |
| time.sleep(max(wait ** attempt, 40)) |
|
|
| return None |
| |
| def hot_fix_task_name(task_name: str): |
| """ |
| This is a hot fix as Hynek inocrrectly named the average collumns |
| """ |
|
|
| if task_name.endswith(":_average"): |
| return task_name.replace(":_average", ":_average|0") |
| return task_name |
|
|
|
|
|
|
|
|
| def load_run_file(run_info: tuple[str, str, list[tuple[datetime, str]]]): |
| run_to_fetch, step, file_data = run_info |
| aggregated_data = {} |
| latest_timestamps = {} |
|
|
| for timestamp, result_file in file_data: |
| file_data = get_file_with_retry(data_folder, result_file) |
| if not file_data: |
| raise Exception(f"File {result_file} not found") |
| json_data = json.loads(file_data) |
| for task, res in json_data["results"].items(): |
| if task not in latest_timestamps or timestamp > latest_timestamps[task]: |
| latest_timestamps[task] = timestamp |
| |
| hashes = json_data["summary_tasks"].get(task, {}).get("hashes") or { |
| "hash_examples": "", |
| "hash_full_prompts": "", |
| "hash_input_tokens": "", |
| "hash_cont_tokens": "" |
| } |
| aggregated_data[task] = { |
| "metrics": res, |
| "hashes": hashes, |
| "filename": result_file |
| } |
|
|
| runname, seed = get_run_name_seed(run_to_fetch) |
| return RunInfo(runname, seed, int(step), |
| [TaskInfo(res["filename"], hot_fix_task_name(task), res["metrics"], res["hashes"]) for task, res in |
| aggregated_data.items()]) |
|
|
| with ThreadPoolExecutor() as pool: |
| run_files = list(itertools.chain.from_iterable( |
| progress.tqdm(pool.map(fetch_run_files, runs_to_fetch), total=len(runs_to_fetch), |
| desc="Fetching datafiles..."))) |
| run_data = list( |
| progress.tqdm(pool.map(load_run_file, run_files), total=len(run_files), desc="Loading evals data...")) |
|
|
| return run_data, None |
|
|
| def filter_run_list_for_language(all_runs, language): |
| if not language: |
| return [] |
| return [ |
| x for x in all_runs if f"-{language}-" in x |
| ] |
|
|
|
|
| def fetch_run_list(results_uri, oauth_token: gr.OAuthToken | None = None, language=None): |
| token = os.environ.get(FALLBACK_TOKEN_NAME) |
| if oauth_token: |
| token = oauth_token.token |
|
|
| data_folder = DataFolder(results_uri, token=token) |
| |
| list_of_runs = [f.removeprefix("results/") for f in |
| data_folder.list_files(subdirectory="results", recursive=False, include_directories=True) |
| if f != "results"] |
| return list_of_runs, gr.update(choices=filter_run_list_for_language(list_of_runs, language), value=None) |
|
|
|
|
| def select_runs_by_regex(runs, current_selected, regex_to_select, lang=None): |
| if lang and "{lang}" in regex_to_select: |
| regex_to_select = regex_to_select.replace("{lang}", lang) |
| comp_re = re.compile(regex_to_select) |
| return list(sorted(set((current_selected if current_selected else []) + |
| [run for run in runs if comp_re.fullmatch(run)]))) |
|
|
|
|
| def add_baseline_rows(df: pd.DataFrame, baseline_runs: list[str], grouping_mode: BASELINE_GROUPING_MODE, baseline_name: str = BASLINE_RUN_NAME) -> pd.DataFrame: |
| if len(baseline_runs) == 0: |
| return df |
| |
| baseline_df = select_runs(df, runs_to_include=baseline_runs) |
| baseline_values = get_groupped_score(baseline_df, baseline_runs, grouping_mode) |
| |
| |
| unique_steps = df["steps"].unique() |
| baseline_values = baseline_values.set_index("steps").reindex(index=unique_steps).interpolate().reset_index() |
| runname, seed = get_run_name_seed(baseline_name) |
|
|
| baseline_values['runname'] = runname |
| baseline_values['seed'] = seed |
| |
| |
| df = pd.concat([df, baseline_values], ignore_index=True) |
|
|
| return df |
|
|
| def normalize_scores(df: pd.DataFrame, normalization_runs: list[str], clip_scores: bool, normalization_mode: NormalizationMode, variability_window: int = 1): |
| if not normalization_runs or len(normalization_runs) == 0: |
| return df |
|
|
| cols_to_process = [col for col in df.columns if is_task_column(col) and not col.endswith('_stderr') and ":_average|" not in col] |
|
|
| if normalization_mode == "Z-norm": |
| df = z_score_normalize(df, normalization_runs, cols_to_process, variability_window) |
| elif normalization_mode == "Rescale": |
| df = rescale_scores(df, normalization_runs, cols_to_process) |
|
|
| if clip_scores: |
| df[cols_to_process] = df[cols_to_process].clip(lower=0) |
| return df |
|
|
| def recompute_averages(df: pd.DataFrame) -> pd.DataFrame: |
| average_columns = [col for col in df.columns if ":_average|" in col] |
| tasks_with_averages = set(x.split(":_average|")[0] for x in average_columns) |
| values_to_average = defaultdict(list) |
| for col in df.columns: |
| if (task := col.split(":")[0]) in tasks_with_averages and (task_subset := col.split(":")[1].split("|")[0]) and task_subset != "_average": |
| task_group = custom_task_aggregate_groups.get(task) |
| |
| if not task_group or task_subset in task_group: |
| values_to_average[(task, col.split("|")[-1])].append(col) |
| |
| for (task, metric), cols in values_to_average.items(): |
| df[f"{task}:_average|{metric}"] = df[cols].mean(axis=1) |
|
|
| return df |
|
|
|
|
| def select_runs_by_language(runs, current_selected, language, selected_cols, mcq_type): |
| new_runs = current_selected |
| if language: |
| if language in agg_score_metrics[mcq_type]: |
| selected_cols = agg_score_metrics[mcq_type][language] |
| new_runs = select_runs_by_regex(runs, current_selected, ".*gemma.*-(" + "|".join(REFERENCE_RUNS) + ")-{lang}-.*", language) |
| return gr.update(value=new_runs, choices=filter_run_list_for_language(runs, language)), gr.update(value=selected_cols if selected_cols else [], choices=selected_cols if selected_cols else []) |
|
|
|
|
| def step_element_match(step_to_check, step_element): |
| step_element = step_element.strip().replace(" ", "") |
| if "-" in step_element: |
| a, b = step_element.split("-") |
| c = None |
| if "%" in b: |
| b, c = b.split("%") |
| return (int(a) <= step_to_check <= int(b) and |
| (c is None or (step_to_check - int(a)) % int(c) == 0)) |
| elif "%" in step_element: |
| return step_to_check % int(step_element[1:]) == 0 |
| else: |
| return step_to_check == int(step_element) |
|
|
|
|
| def init_input_component_values(run_data: RunData, normalization_mode: NormalizationMode, language: str | None = None): |
| task_metrics = set(metric for run in run_data for task in run.tasks for metric in task.metrics.keys()) |
| initial_value = "agg_score_metrics" if language and language in agg_score_metrics_dict_prob else \ |
| ("acc_norm" if "acc_norm" in task_metrics else next(iter(task_metrics), None)) |
| runs = set(run.full_name for run in run_data) |
| baseline_runs = [run for run in runs if is_baseline_run(run)] |
|
|
|
|
| |
|
|
| return (gr.update(choices=["agg_score_metrics"] + sorted(task_metrics, key=lambda m: (m.endswith("_stderr"), m)), value=[initial_value]), |
| init_input_normalization_runs(run_data, normalization_mode), |
| gr.update(value=[] if not baseline_runs else [baseline_runs[0]], choices=sorted(runs))) |
|
|
|
|
| def init_input_normalization_runs(runs: RunData, normalization_mode: NormalizationMode): |
| run_names = set([run.full_name for run in runs]) |
| if normalization_mode == "Z-norm": |
| referene_runs = [run for run in run_names if is_reference_run(run)] |
| return gr.update(value=referene_runs, choices=sorted(run_names)) |
| |
| elif normalization_mode == "Rescale": |
| baseline_runs = [run for run in run_names if is_baseline_run(run)] |
| return gr.update(value=baseline_runs, choices=sorted(run_names)) |
| |
| else: |
| return gr.update(value=[], choices=[]) |
|
|
|
|
|
|
| def init_std_dev_runs(runs, current_val): |
| |
| value = current_val or "-" |
| seed_counter = Counter() |
| for run in runs or []: |
| seed_counter[run.split("-seed-")[0]] += 1 |
|
|
| if seed_counter[value] <= 1: |
| top_val, top_count = seed_counter.most_common(n=1)[0] if seed_counter else (None, 0) |
| value = top_val if top_count > 1 else "-" |
| return gr.update(value=value, choices=["-"] + sorted([val for val, count in seed_counter.items() if count > 1])) |
|
|
|
|
| def update_dropdown_choices(selected_choices, possible_choices): |
| selected_choices = [choice for choice in selected_choices if choice in possible_choices] |
| return gr.update(choices=possible_choices, value=selected_choices) |
|
|
|
|
| def render_results_table(df: pd.DataFrame, metrics, task_avg, normalization_runs: list[str], baseline_runs: list[str], baseline_mode: BASELINE_GROUPING_MODE, clip_scores: bool, |
| normalization_mode: NormalizationMode, aggregate_score_cols: list[str], language: str, variability_window: int = 1, mcq_type = "prob"): |
| |
| |
| |
|
|
| |
| df = add_baseline_rows(df, baseline_runs, baseline_mode) |
|
|
| |
| df = normalize_scores(df, normalization_runs=normalization_runs, clip_scores=clip_scores, normalization_mode=normalization_mode, variability_window=variability_window) |
|
|
| df = recompute_averages(df) |
| |
| |
| df = select_runs(df, runs_to_exclude=baseline_runs) |
|
|
| to_drop = [] |
| for col in df.columns: |
| if is_task_column(col): |
| |
| if "agg_score_metrics" in metrics and language in agg_score_metrics[mcq_type] and col in agg_score_metrics[mcq_type][language]: |
| continue |
| task, metric = col.split("/") |
| |
| if ((metric not in metrics and len(metrics) > 0) or |
| (":_average|" in task and "show averages" not in task_avg) or |
| ("|" in task and ":_average|" not in task and ":" in task.split("|")[ |
| 1] and "show expanded" not in task_avg)): |
| to_drop.append(col) |
| if to_drop: |
| df = df.drop(columns=to_drop) |
|
|
|
|
| df.sort_values(by=["runname", "seed", "steps"], inplace=True) |
| df = update_agg_score(df, aggregate_score_cols) |
| |
| aggregate_columns = [col for col in df.columns if is_aggregate_column(col)] |
| |
| task_columns = [col for col in df.columns if is_task_column(col)] |
| return df, update_dropdown_choices(aggregate_score_cols, task_columns), gr.update(aggregate_columns[0], choices=aggregate_columns + task_columns) |
|
|
|
|
| def get_type_tasks_dict(tasks: list[str]) -> dict[TASK_TYPE, list[str]]: |
| """ |
| Creates a dictionary mapping task types to lists of task names. |
| |
| Args: |
| tasks (list[str]): List of task names. |
| |
| Returns: |
| dict[TASK_TYPE, list[str]]: Dictionary with task types as keys and lists of task names as values. |
| """ |
| task_type_dict: dict[TASK_TYPE, list[str]] = defaultdict(list) |
| |
| for task in tasks: |
| task_type = get_task_type(task) |
| if not task_type: |
| raise ValueError(f"Task {task} has no task type") |
| task_type_dict[task_type].append(task) |
| |
| return task_type_dict |
|
|
| def update_agg_score(df: pd.DataFrame, agg_score_columns: list[str]) -> pd.DataFrame: |
| if not agg_score_columns or df is None or "steps" not in df: |
| return df |
|
|
| new_df = df.copy() |
| cols_to_avg = [col for col in agg_score_columns if col in new_df.columns] |
| |
| if cols_to_avg: |
| |
| task_type_dict = get_type_tasks_dict(cols_to_avg) |
| |
| for task_type, tasks in task_type_dict.items(): |
| new_df[f'agg_score_{task_type}'] = new_df[tasks].mean(axis=1) |
| |
|
|
| |
| new_df['agg_score_macro'] = new_df[[f'agg_score_{task_type}' for task_type in task_type_dict.keys()]].mean(axis=1) |
|
|
| |
| new_df['agg_score_micro'] = new_df[cols_to_avg].mean(axis=1) |
|
|
| return new_df |
|
|
|
|
| def export_results_csv(df): |
| df.to_csv("output.csv", index=False) |
| return gr.update(value="output.csv", visible=True) |
|
|
|
|
| def check_missing_datapoints(runs, steps_to_check, run_data: RunData, check_missing_checkpoints): |
| if not runs or check_missing_checkpoints == CHECK_MISSING_DATAPOINTS_BUTTON_CLOSE_LABEL or not run_data or not steps_to_check: |
| return gr.Json(value={}, visible=False), gr.Button(value=CHECK_MISSING_DATAPOINTS_BUTTON_LABEL) |
|
|
| max_step = max(run.step for run in run_data) |
| steps_set = set() |
| for step_elem in steps_to_check.split(","): |
| step_element = step_elem.strip().replace(" ", "") |
| if "-" in step_element: |
| a, b = step_element.split("-") |
| c = None |
| if "%" in b: |
| b, c = b.split("%") |
| steps_set.update(range(int(a), int(b) + 1, int(c) if c else 1)) |
| elif "%" in step_element: |
| steps_set.update(range(0, max_step + 1, int(step_element[1:]))) |
| else: |
| steps_set.add(int(step_element)) |
|
|
| existing_evals = {(run.name, run.seed, run.step) for run in run_data} |
|
|
| missing_evals = defaultdict(dict) |
| for run in runs: |
| runname, seed = get_run_name_seed(run) |
| missing_steps = [ |
| step for step in sorted(steps_set) if (runname, seed, step) not in existing_evals |
| ] |
| if missing_steps: |
| missing_evals[runname][str(seed)] = missing_steps |
|
|
| return gr.Json(value=missing_evals, visible=True), gr.Button(value=CHECK_MISSING_DATAPOINTS_BUTTON_CLOSE_LABEL) |
|
|