File size: 7,462 Bytes
5e4510c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
#!/usr/bin/env python3
"""
Random Baseline for K-Module Problem

This script establishes a baseline by randomly sampling configurations.
It simulates what you'd get with pass@N (N independent random attempts)
without any learning or optimization.

This is useful because:
1. It establishes the "no learning" baseline
2. For closed models that don't support n>1 responses, we can't do true pass@k
3. Shows the expected performance of random search

Usage:
    python run_random_baseline.py [--samples 100] [--trials 10]
"""

import argparse
import json
import random
import time
from pathlib import Path

# Import the evaluator
from evaluator import VALID_OPTIONS, CORRECT_CONFIG, NUM_MODULES


def generate_random_config() -> dict:
    """Generate a random pipeline configuration."""
    return {
        module: random.choice(options)
        for module, options in VALID_OPTIONS.items()
    }


def score_config(config: dict) -> int:
    """Score a configuration (number of correct modules)."""
    return sum(
        1 for module, value in config.items()
        if CORRECT_CONFIG.get(module) == value
    )


def run_random_search(max_samples: int) -> dict:
    """
    Run random search until solution found or max_samples reached.

    Returns:
        dict with results
    """
    results = {
        "samples": [],
        "scores": [],
        "best_scores": [],
        "solution_found_at": None,
        "configs_tried": [],
    }

    best_so_far = 0

    for i in range(max_samples):
        config = generate_random_config()
        score = score_config(config)

        results["samples"].append(i)
        results["scores"].append(score / NUM_MODULES)
        best_so_far = max(best_so_far, score)
        results["best_scores"].append(best_so_far / NUM_MODULES)
        results["configs_tried"].append(config)

        if score == NUM_MODULES and results["solution_found_at"] is None:
            results["solution_found_at"] = i

    return results


def run_multiple_trials(num_trials: int, max_samples: int) -> list:
    """Run multiple independent trials of random search."""
    trial_results = []

    for trial in range(num_trials):
        random.seed(trial * 1000 + int(time.time()))  # Different seed per trial
        result = run_random_search(max_samples)
        trial_results.append({
            "trial": trial,
            "solution_found_at": result["solution_found_at"],
            "final_best_score": result["best_scores"][-1] if result["best_scores"] else 0,
            "scores": result["scores"],
            "best_scores": result["best_scores"],
        })

    return trial_results


def calculate_pass_at_k(trial_results: list, k_values: list) -> dict:
    """
    Calculate pass@k metrics.

    pass@k = probability of finding solution within k samples

    For random search on 625 possibilities:
    - pass@1 = 1/625 = 0.16%
    - pass@100 ≈ 1 - (624/625)^100 ≈ 14.8%
    - pass@312 ≈ 50% (half the search space)
    """
    pass_at_k = {}

    for k in k_values:
        successes = sum(
            1 for r in trial_results
            if r["solution_found_at"] is not None and r["solution_found_at"] < k
        )
        pass_at_k[k] = successes / len(trial_results)

    return pass_at_k


def theoretical_pass_at_k(k: int, search_space: int = 625) -> float:
    """Calculate theoretical pass@k for uniform random search."""
    # Probability of NOT finding solution in k tries
    prob_fail = ((search_space - 1) / search_space) ** k
    return 1 - prob_fail


def main():
    parser = argparse.ArgumentParser(description="Random baseline for K-Module problem")
    parser.add_argument("--samples", type=int, default=100, help="Max samples per trial")
    parser.add_argument("--trials", type=int, default=100, help="Number of independent trials")
    parser.add_argument("--output", default="random_baseline_output", help="Output directory")
    args = parser.parse_args()

    print("=" * 60)
    print("K-MODULE PROBLEM: RANDOM BASELINE")
    print("=" * 60)
    print(f"Search space: {5**NUM_MODULES} configurations")
    print(f"Running {args.trials} trials with up to {args.samples} samples each")
    print()

    # Run trials
    print("Running random search trials...")
    trial_results = run_multiple_trials(args.trials, args.samples)

    # Calculate statistics
    solutions_found = [r for r in trial_results if r["solution_found_at"] is not None]
    success_rate = len(solutions_found) / len(trial_results)

    if solutions_found:
        avg_samples_to_solution = sum(r["solution_found_at"] for r in solutions_found) / len(solutions_found)
        min_samples = min(r["solution_found_at"] for r in solutions_found)
        max_samples = max(r["solution_found_at"] for r in solutions_found)
    else:
        avg_samples_to_solution = float('inf')
        min_samples = max_samples = None

    # Calculate pass@k
    k_values = [1, 10, 20, 50, 100, 200, 312]
    k_values = [k for k in k_values if k <= args.samples]
    empirical_pass_at_k = calculate_pass_at_k(trial_results, k_values)

    # Print results
    print("\n### Results")
    print(f"  Success rate: {success_rate:.1%} ({len(solutions_found)}/{len(trial_results)} trials)")
    if solutions_found:
        print(f"  Avg samples to solution: {avg_samples_to_solution:.1f}")
        print(f"  Min samples: {min_samples}")
        print(f"  Max samples: {max_samples}")
    else:
        print(f"  No solutions found in {args.samples} samples")

    print("\n### Pass@k Comparison (Empirical vs Theoretical)")
    print("  k     | Empirical | Theoretical")
    print("  ------|-----------|------------")
    for k in k_values:
        emp = empirical_pass_at_k.get(k, 0)
        theo = theoretical_pass_at_k(k)
        print(f"  {k:5d} | {emp:8.1%}  | {theo:8.1%}")

    # Save results
    output_dir = Path(args.output)
    output_dir.mkdir(exist_ok=True)

    results = {
        "config": {
            "samples_per_trial": args.samples,
            "num_trials": args.trials,
            "search_space": 5 ** NUM_MODULES,
        },
        "summary": {
            "success_rate": success_rate,
            "avg_samples_to_solution": avg_samples_to_solution if solutions_found else None,
            "min_samples": min_samples,
            "max_samples": max_samples,
        },
        "pass_at_k": {
            "empirical": empirical_pass_at_k,
            "theoretical": {k: theoretical_pass_at_k(k) for k in k_values},
        },
        "trials": trial_results,
    }

    with open(output_dir / "random_baseline_results.json", "w") as f:
        json.dump(results, f, indent=2)

    print(f"\nResults saved to: {output_dir}/random_baseline_results.json")

    # Key insight
    print("\n### Key Insight")
    print("  Random search requires ~312 samples (50% of search space) on average.")
    print("  This is the baseline that any optimization method should beat.")
    print("  ")
    print("  For LLM-based methods:")
    print("  - pass@k with closed models requires k separate API calls")
    print("  - Each call is independent (no learning across calls)")
    print("  - This is equivalent to random search if prompts don't help")
    print("  ")
    print("  OpenEvolve should find solutions in <<312 evaluations by:")
    print("  - Learning from population diversity")
    print("  - Combining good 'building blocks' via crossover")

    print("\n" + "=" * 60)


if __name__ == "__main__":
    main()