|
|
import argparse |
|
|
import json |
|
|
from pathlib import Path |
|
|
from typing import Dict, List, Tuple |
|
|
|
|
|
import numpy as np |
|
|
import pandas as pd |
|
|
import yaml |
|
|
from omegaconf import DictConfig, OmegaConf |
|
|
from tqdm import tqdm |
|
|
from metrics import compute_metrics |
|
|
|
|
|
|
|
|
def extract_config_from_log(log_path: Path) -> Dict: |
|
|
""" |
|
|
Parse the experiment configuration from run_experiment.log file. |
|
|
The config is in YAML format at the beginning of the log file. |
|
|
""" |
|
|
with open(log_path, 'r', encoding="latin1") as f: |
|
|
lines = f.readlines() |
|
|
|
|
|
|
|
|
yaml_lines = [] |
|
|
in_yaml = False |
|
|
|
|
|
for line in lines: |
|
|
|
|
|
if not in_yaml and line.strip() and not line.startswith('['): |
|
|
in_yaml = True |
|
|
|
|
|
|
|
|
if in_yaml and line.startswith('['): |
|
|
break |
|
|
|
|
|
if in_yaml: |
|
|
yaml_lines.append(line) |
|
|
|
|
|
|
|
|
yaml_content = ''.join(yaml_lines) |
|
|
config_dict = yaml.safe_load(yaml_content) |
|
|
|
|
|
|
|
|
return OmegaConf.create(config_dict) |
|
|
|
|
|
|
|
|
def load_inference_results_by_grader(jsonl_path: Path) -> Tuple[Dict[str, Dict], Dict[str, Dict]]: |
|
|
""" |
|
|
Load predictions and labels from the inference results JSONL file, organized by grader. |
|
|
|
|
|
Returns: |
|
|
grader_a_data: Dictionary mapping essay_id to {'prediction': score, 'label': label} |
|
|
grader_b_data: Dictionary mapping essay_id to {'prediction': score, 'label': label} |
|
|
""" |
|
|
grader_a_data = {} |
|
|
grader_b_data = {} |
|
|
|
|
|
with open(jsonl_path, 'r', encoding='utf-8') as f: |
|
|
for line in f: |
|
|
data = json.loads(line.strip()) |
|
|
essay_id = (data['id'], data['id_prompt'], data['essay_text']) |
|
|
|
|
|
model_types = ["gpt", "sabia", "deepseek"] |
|
|
prediction_field = "pontuacao" if any(model in jsonl_path.name for model in model_types) else "prediction" |
|
|
prediction = data[prediction_field] |
|
|
essay_data = { |
|
|
'prediction': prediction, |
|
|
'label': data['label'] |
|
|
} |
|
|
|
|
|
if data['reference'] == 'grader_a': |
|
|
grader_a_data[essay_id] = essay_data |
|
|
elif data['reference'] == 'grader_b': |
|
|
grader_b_data[essay_id] = essay_data |
|
|
assert len(grader_a_data) == len(grader_b_data), "Mismatch in number of essays graded by A and B" |
|
|
return grader_a_data, grader_b_data |
|
|
|
|
|
|
|
|
def compute_bootstrap_confidence_intervals_two_graders( |
|
|
grader_a_data: Dict[str, Dict], |
|
|
grader_b_data: Dict[str, Dict], |
|
|
metrics_to_compute: List[str], |
|
|
cfg: DictConfig, |
|
|
n_bootstrap: int = 1000, |
|
|
confidence_level: float = 0.95, |
|
|
random_state: int = 42, |
|
|
) -> Dict[str, Tuple[float, float, float]]: |
|
|
""" |
|
|
Compute bootstrap confidence intervals for specified metrics using two-grader structure. |
|
|
|
|
|
For each bootstrap sample: |
|
|
1. Sample essay IDs with replacement |
|
|
2. For each sampled essay ID, get both grader A and grader B predictions/labels |
|
|
3. Compute metrics separately for grader A and grader B |
|
|
4. Take the mean of the two grader metrics |
|
|
|
|
|
Parameters: |
|
|
grader_a_data: Dictionary mapping essay_id to prediction/label for grader A |
|
|
grader_b_data: Dictionary mapping essay_id to prediction/label for grader B |
|
|
metrics_to_compute: List of metric names to compute CIs for |
|
|
cfg: Configuration object |
|
|
n_bootstrap: Number of bootstrap samples |
|
|
confidence_level: Confidence level (default 0.95 for 95% CI) |
|
|
random_state: Random seed for reproducibility |
|
|
|
|
|
Returns: |
|
|
Dictionary mapping metric names to (mean, lower_bound, upper_bound) |
|
|
""" |
|
|
if random_state is not None: |
|
|
np.random.seed(random_state) |
|
|
|
|
|
|
|
|
essay_ids = list(grader_a_data.keys()) |
|
|
assert set(essay_ids) == set(grader_b_data.keys()), "Essay IDs don't match between graders" |
|
|
|
|
|
n_essays = len(essay_ids) |
|
|
bootstrap_metrics = {metric: [] for metric in metrics_to_compute} |
|
|
|
|
|
|
|
|
for _ in tqdm(range(n_bootstrap), desc="Performing Bootstrap samples"): |
|
|
|
|
|
sampled_indices = np.random.choice(n_essays, size=n_essays, replace=True) |
|
|
|
|
|
|
|
|
grader_a_predictions = [] |
|
|
grader_a_labels = [] |
|
|
grader_b_predictions = [] |
|
|
grader_b_labels = [] |
|
|
|
|
|
for idx in sampled_indices: |
|
|
essay_id = essay_ids[idx] |
|
|
grader_a_predictions.append(grader_a_data[essay_id]['prediction']) |
|
|
grader_a_labels.append(grader_a_data[essay_id]['label']) |
|
|
grader_b_predictions.append(grader_b_data[essay_id]['prediction']) |
|
|
grader_b_labels.append(grader_b_data[essay_id]['label']) |
|
|
|
|
|
|
|
|
grader_a_predictions = np.array(grader_a_predictions) |
|
|
grader_a_labels = np.array(grader_a_labels) |
|
|
grader_b_predictions = np.array(grader_b_predictions) |
|
|
grader_b_labels = np.array(grader_b_labels) |
|
|
|
|
|
|
|
|
metrics_a = compute_metrics((grader_a_predictions, grader_a_labels), cfg) |
|
|
metrics_b = compute_metrics((grader_b_predictions, grader_b_labels), cfg) |
|
|
|
|
|
|
|
|
for metric in metrics_to_compute: |
|
|
if metric in metrics_a and metric in metrics_b: |
|
|
mean_metric = (metrics_a[metric] + metrics_b[metric]) / 2 |
|
|
bootstrap_metrics[metric].append(mean_metric) |
|
|
|
|
|
|
|
|
alpha = 1 - confidence_level |
|
|
lower_percentile = (alpha / 2) * 100 |
|
|
upper_percentile = (1 - alpha / 2) * 100 |
|
|
|
|
|
ci_results = {} |
|
|
for metric, values in bootstrap_metrics.items(): |
|
|
if values: |
|
|
values_array = np.array(values) |
|
|
mean_val = np.mean(values_array) |
|
|
lower_bound = np.percentile(values_array, lower_percentile) |
|
|
upper_bound = np.percentile(values_array, upper_percentile) |
|
|
ci_results[metric] = (mean_val, lower_bound, upper_bound) |
|
|
|
|
|
return ci_results |
|
|
|
|
|
|
|
|
def save_results_to_csv( |
|
|
experiment_id: str, |
|
|
ci_results: Dict[str, Tuple[float, float, float]], |
|
|
output_path: Path, |
|
|
) -> None: |
|
|
""" |
|
|
Save bootstrap CI results to CSV file with flattened structure. |
|
|
""" |
|
|
from datetime import datetime |
|
|
|
|
|
|
|
|
row_data = { |
|
|
'experiment_id': experiment_id, |
|
|
'timestamp': datetime.now().strftime("%Y-%m-%d %H:%M:%S") |
|
|
} |
|
|
|
|
|
|
|
|
for metric, (mean_val, lower, upper) in ci_results.items(): |
|
|
row_data[f'{metric}_mean'] = mean_val |
|
|
row_data[f'{metric}_lower_95ci'] = lower |
|
|
row_data[f'{metric}_upper_95ci'] = upper |
|
|
row_data[f'{metric}_ci_width'] = upper - lower |
|
|
|
|
|
|
|
|
df = pd.DataFrame([row_data]) |
|
|
|
|
|
|
|
|
df.to_csv(output_path, index=False) |
|
|
print(f"Results saved to {output_path}") |
|
|
|
|
|
|
|
|
def main(): |
|
|
""" |
|
|
# Basic usage |
|
|
python compute_bootstrap_ci.py /path/to/experiment/directory |
|
|
|
|
|
# With custom parameters |
|
|
python compute_bootstrap_ci.py /path/to/experiment/directory \ |
|
|
--metrics QWK accuracy RMSE HDIV \ |
|
|
--n-bootstrap 2000 \ |
|
|
--confidence-level 0.99 |
|
|
""" |
|
|
parser = argparse.ArgumentParser( |
|
|
description='Compute bootstrap confidence intervals for API model inference results with two-grader structure' |
|
|
) |
|
|
parser.add_argument( |
|
|
'experiment_dir', |
|
|
type=str, |
|
|
help='Path to the experiment directory containing log and results files' |
|
|
) |
|
|
parser.add_argument( |
|
|
'--metrics', |
|
|
nargs='+', |
|
|
default=['QWK', 'Macro_F1', 'Weighted_F1'], |
|
|
help='Metrics to compute confidence intervals for (default: QWK Macro_F1 Weighted_F1)' |
|
|
) |
|
|
parser.add_argument( |
|
|
'--n-bootstrap', |
|
|
type=int, |
|
|
default=1000, |
|
|
help='Number of bootstrap samples (default: 1000)' |
|
|
) |
|
|
parser.add_argument( |
|
|
'--confidence-level', |
|
|
type=float, |
|
|
default=0.95, |
|
|
help='Confidence level for intervals (default: 0.95)' |
|
|
) |
|
|
parser.add_argument( |
|
|
'--seed', |
|
|
type=int, |
|
|
default=42, |
|
|
help='Random seed for reproducibility (default: 42)' |
|
|
) |
|
|
|
|
|
args = parser.parse_args() |
|
|
|
|
|
|
|
|
exp_dir = Path(args.experiment_dir) |
|
|
if not exp_dir.exists(): |
|
|
raise FileNotFoundError(f"Experiment directory not found: {exp_dir}") |
|
|
|
|
|
|
|
|
log_files = list(exp_dir.glob("*run_inference_experiment.log")) |
|
|
if not log_files: |
|
|
raise FileNotFoundError(f"No log file found in {exp_dir}") |
|
|
log_path = log_files[0] |
|
|
|
|
|
|
|
|
results_files = list(exp_dir.glob("*_inference_results.jsonl")) |
|
|
if not results_files: |
|
|
raise FileNotFoundError(f"No inference results file found in {exp_dir}") |
|
|
results_path = results_files[0] |
|
|
|
|
|
|
|
|
experiment_id = exp_dir.name |
|
|
|
|
|
print(f"Processing experiment: {experiment_id}") |
|
|
print(f"Log file: {log_path}") |
|
|
print(f"Results file: {results_path}") |
|
|
|
|
|
|
|
|
try: |
|
|
config_dict = extract_config_from_log(log_path) |
|
|
|
|
|
cfg = OmegaConf.create(config_dict) |
|
|
seed = cfg.experiments.get('training_params', {}).get('seed', args.seed) |
|
|
except Exception as e: |
|
|
raise RuntimeError(f"Failed to extract configuration from log file: {e}") |
|
|
|
|
|
|
|
|
grader_a_data, grader_b_data = load_inference_results_by_grader(results_path) |
|
|
print(f"Loaded {len(grader_a_data)} essays with data from both graders") |
|
|
|
|
|
|
|
|
ci_results = compute_bootstrap_confidence_intervals_two_graders( |
|
|
grader_a_data=grader_a_data, |
|
|
grader_b_data=grader_b_data, |
|
|
metrics_to_compute=args.metrics, |
|
|
cfg=cfg, |
|
|
n_bootstrap=args.n_bootstrap, |
|
|
confidence_level=args.confidence_level, |
|
|
random_state=seed, |
|
|
) |
|
|
|
|
|
|
|
|
all_predictions_a = np.array([data['prediction'] for data in grader_a_data.values()]) |
|
|
all_labels_a = np.array([data['label'] for data in grader_a_data.values()]) |
|
|
all_predictions_b = np.array([data['prediction'] for data in grader_b_data.values()]) |
|
|
all_labels_b = np.array([data['label'] for data in grader_b_data.values()]) |
|
|
|
|
|
metrics_full_a = compute_metrics((all_predictions_a, all_labels_a), cfg) |
|
|
metrics_full_b = compute_metrics((all_predictions_b, all_labels_b), cfg) |
|
|
|
|
|
print("\nFull Dataset Metrics:") |
|
|
print(" Grader A:") |
|
|
for metric in args.metrics: |
|
|
if metric in metrics_full_a: |
|
|
print(f" {metric}: {metrics_full_a[metric]:.4f}") |
|
|
print(" Grader B:") |
|
|
for metric in args.metrics: |
|
|
if metric in metrics_full_b: |
|
|
print(f" {metric}: {metrics_full_b[metric]:.4f}") |
|
|
print(" Mean (A+B)/2:") |
|
|
for metric in args.metrics: |
|
|
if metric in metrics_full_a and metric in metrics_full_b: |
|
|
mean_val = (metrics_full_a[metric] + metrics_full_b[metric]) / 2 |
|
|
print(f" {metric}: {mean_val:.4f}") |
|
|
|
|
|
|
|
|
print(f"\nBootstrap Confidence Intervals ({args.confidence_level*100:.0f}%):") |
|
|
print(" (Based on mean of grader A and B metrics)") |
|
|
for metric, (mean_val, lower, upper) in ci_results.items(): |
|
|
print(f" {metric}: {mean_val:.4f} [{lower:.4f}, {upper:.4f}]") |
|
|
|
|
|
|
|
|
output_path = exp_dir / "bootstrap_confidence_intervals.csv" |
|
|
save_results_to_csv(experiment_id, ci_results, output_path) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|