codearena-rl / analyze_finetuning.py
havinashpatil
Add comprehensive LLM finetuning analysis with 7 visualization graphs
9d429ce
"""
Advanced Analysis of LLM Finetuning Performance
Analyzes reward curves, complexity metrics, and fixer method effectiveness
"""
import pandas as pd
import matplotlib.pyplot as plt
import numpy as np
import os
from collections import Counter
os.makedirs('results', exist_ok=True)
# Load data
rewards_df = pd.read_csv('rewards_log.csv')
complexity_df = pd.read_csv('complexity_rewards.csv')
print("\n" + "="*70)
print("FINETUNING ANALYSIS REPORT")
print("="*70)
# ─── SUMMARY STATISTICS ──────────────────────────────────────────────────────
print("\nπŸ“Š TRAINING OVERVIEW")
print(f"Total Episodes: {len(rewards_df)}")
print(f"Unique Tasks: {rewards_df['task_id'].nunique()}")
print(f"Date Range: {rewards_df['timestamp'].iloc[0]} to {rewards_df['timestamp'].iloc[-1]}")
print(f"Avg Reward: {rewards_df['reward'].mean():.4f}")
print(f"Max Reward: {rewards_df['reward'].max():.4f}")
print(f"Min Reward: {rewards_df['reward'].min():.4f}")
print(f"Reward Std: {rewards_df['reward'].std():.4f}")
# ─── TASK BREAKDOWN ──────────────────────────────────────────────────────────
print("\nπŸ“‹ PERFORMANCE BY TASK")
task_stats = rewards_df.groupby('task_id')['reward'].agg([
('Count', 'count'),
('Mean', 'mean'),
('Max', 'max'),
('Min', 'min'),
('Std', 'std')
]).round(4)
print(task_stats)
# ─── COMPLEXITY ANALYSIS ────────────────────────────────────────────────────────
print("\n⚑ COMPLEXITY VS REWARD ANALYSIS")
complexity_stats = complexity_df.groupby('complexity')['reward'].agg([
('Count', 'count'),
('Mean Reward', 'mean'),
('Max Reward', 'max'),
('Min Reward', 'min')
]).round(4)
print(complexity_stats)
# ─── METHOD PERFORMANCE ──────────────────────────────────────────────────────
print("\nπŸ”§ FIXER METHOD EFFECTIVENESS")
method_stats = complexity_df.groupby('method')['reward'].agg([
('Count', 'count'),
('Mean Reward', 'mean'),
('Max Reward', 'max'),
('Min Reward', 'min')
]).round(4)
print(method_stats)
# ─── COMPLEXITY BREAKDOWN ──────────────────────────────────────────────────────
print("\nπŸ”„ COMPLEXITY DISTRIBUTION")
complexity_counts = complexity_df['complexity'].value_counts().sort_values(ascending=False)
print(complexity_counts)
# ─── GRAPH 1: Complexity vs Reward Scatter ──────────────────────────────────────
fig, ax = plt.subplots(figsize=(12, 6))
colors = {'ollama': 'blue', 'builtin': 'red', 'tgi': 'green'}
for method in complexity_df['method'].unique():
df_method = complexity_df[complexity_df['method'] == method]
ax.scatter(range(len(df_method)), df_method['reward'],
label=f"{method.capitalize()} (n={len(df_method)})",
alpha=0.6, s=60, color=colors.get(method, 'gray'))
ax.set_xlabel('Sample Index', fontsize=11)
ax.set_ylabel('Reward Score (0-1)', fontsize=11)
ax.set_title('LLM Fixer Method Performance Comparison', fontsize=13, fontweight='bold')
ax.legend(loc='best')
ax.grid(True, alpha=0.3)
plt.tight_layout()
plt.savefig('results/method_performance.png', dpi=150)
plt.close()
print("\nβœ“ Saved: method_performance.png")
# ─── GRAPH 2: Complexity Distribution (Pie + Bar) ──────────────────────────────
fig, (ax1, ax2) = plt.subplots(1, 2, figsize=(14, 5))
# Pie chart
colors_pie = plt.cm.Set3(np.linspace(0, 1, len(complexity_counts)))
ax1.pie(complexity_counts.values, labels=complexity_counts.index, autopct='%1.1f%%',
colors=colors_pie, startangle=90)
ax1.set_title('Complexity Distribution in Dataset', fontsize=12, fontweight='bold')
# Bar chart
complexity_counts.plot(kind='bar', ax=ax2, color='skyblue', edgecolor='navy', alpha=0.7)
ax2.set_xlabel('Time Complexity Class', fontsize=11)
ax2.set_ylabel('Number of Samples', fontsize=11)
ax2.set_title('Complexity Class Frequency', fontsize=12, fontweight='bold')
ax2.set_xticklabels(ax2.get_xticklabels(), rotation=45)
ax2.grid(axis='y', alpha=0.3)
plt.tight_layout()
plt.savefig('results/complexity_distribution.png', dpi=150)
plt.close()
print("βœ“ Saved: complexity_distribution.png")
# ─── GRAPH 3: Method Performance Box Plot ──────────────────────────────────────
fig, ax = plt.subplots(figsize=(10, 6))
method_data = [complexity_df[complexity_df['method'] == m]['reward'].values
for m in complexity_df['method'].unique()]
bp = ax.boxplot(method_data, labels=complexity_df['method'].unique(), patch_artist=True)
for patch, color in zip(bp['boxes'], ['lightblue', 'lightcoral', 'lightgreen'][:len(bp['boxes'])]):
patch.set_facecolor(color)
ax.set_xlabel('Fixer Method', fontsize=11)
ax.set_ylabel('Reward Score (0-1)', fontsize=11)
ax.set_title('Reward Distribution by Fixer Method', fontsize=13, fontweight='bold')
ax.grid(axis='y', alpha=0.3)
plt.tight_layout()
plt.savefig('results/method_boxplot.png', dpi=150)
plt.close()
print("βœ“ Saved: method_boxplot.png")
# ─── GRAPH 4: Task Performance Heatmap ──────────────────────────────────────────
task_reward_matrix = rewards_df.pivot_table(
values='reward',
index='task_id',
aggfunc=['mean', 'max', 'std']
)
task_reward_matrix = task_reward_matrix.droplevel(0, axis=1)
fig, ax = plt.subplots(figsize=(10, 6))
im = ax.imshow(task_reward_matrix.values, cmap='RdYlGn', aspect='auto', vmin=0, vmax=1)
ax.set_xticks(range(len(task_reward_matrix.columns)))
ax.set_yticks(range(len(task_reward_matrix.index)))
ax.set_xticklabels(task_reward_matrix.columns, rotation=45)
ax.set_yticklabels(task_reward_matrix.index)
ax.set_title('Task Difficulty Performance Matrix (Mean, Max, Std)', fontsize=13, fontweight='bold')
# Add text annotations
for i in range(len(task_reward_matrix.index)):
for j in range(len(task_reward_matrix.columns)):
text = ax.text(j, i, f'{task_reward_matrix.values[i, j]:.2f}',
ha="center", va="center", color="black", fontsize=9)
plt.colorbar(im, ax=ax, label='Reward Score')
plt.tight_layout()
plt.savefig('results/task_performance_matrix.png', dpi=150)
plt.close()
print("βœ“ Saved: task_performance_matrix.png")
# ─── GRAPH 5: Cumulative Reward Over Time ──────────────────────────────────────
fig, ax = plt.subplots(figsize=(12, 6))
sorted_rewards = complexity_df.sort_values('timestamp')
cumulative_reward = sorted_rewards['reward'].cumsum()
ax.plot(range(len(cumulative_reward)), cumulative_reward, marker='o',
markersize=4, linewidth=2, color='darkblue', alpha=0.7, label='Cumulative Reward')
ax.fill_between(range(len(cumulative_reward)), cumulative_reward, alpha=0.2, color='blue')
ax.set_xlabel('Sample Index (Chronological)', fontsize=11)
ax.set_ylabel('Cumulative Reward', fontsize=11)
ax.set_title('Cumulative Reward Trajectory', fontsize=13, fontweight='bold')
ax.grid(True, alpha=0.3)
ax.legend()
plt.tight_layout()
plt.savefig('results/cumulative_reward.png', dpi=150)
plt.close()
print("βœ“ Saved: cumulative_reward.png")
# ─── FINAL SUMMARY ──────────────────────────────────────────────────────────────
print("\n" + "="*70)
print("βœ… ALL GRAPHS GENERATED IN results/ DIRECTORY:")
print(" β€’ reward_curve.png (rolling avg of rewards)")
print(" β€’ reward_by_task.png (task-wise comparison)")
print(" β€’ method_performance.png (fixer methods)")
print(" β€’ complexity_distribution.png (algorithm classes)")
print(" β€’ method_boxplot.png (reward distribution)")
print(" β€’ task_performance_matrix.png (heatmap)")
print(" β€’ cumulative_reward.png (training trajectory)")
print("="*70 + "\n")