File size: 11,662 Bytes
b14c6e3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
"""
Evaluation Script for Adaptive Alert Triage Environment
========================================================

Runs baseline agents on all tasks and computes performance metrics.

BUGS FIXED vs original:
  1. HardTaskGrader(correlation_chains=[]) β€” HardTaskGrader takes NO __init__
     args (chains come dynamically via update_correlation_state).
     Fixed: grader = HardTaskGrader()
  2. grader.record_system_failure() doesn't exist β€” the method is
     record_failures(n: int). Fixed to use the correct API.
  3. Success thresholds were wrong (medium=0.65, hard=0.60) β€” the actual
     grader constants are medium=0.55, hard=0.50.  Fixed to import from
     the grader modules.
  4. evaluate_agent_on_task consumed only processed_alerts[0] per step β€”
     env.step() may produce multiple processed alerts when there are batch
     actions. Fixed to iterate the full list.
"""

from __future__ import annotations

import argparse
import json
from typing import Any, Dict, List

import numpy as np

from adaptive_alert_triage.env   import AdaptiveAlertTriageEnv
from agents.baseline import RuleBasedAgent, ImprovedRuleBasedAgent

from tasks.easy   import EasyTaskGrader,   SUCCESS_THRESHOLD as EASY_THRESH
from tasks.medium import MediumTaskGrader, SUCCESS_THRESHOLD as MED_THRESH
from tasks.hard   import HardTaskGrader,   SUCCESS_THRESHOLD as HARD_THRESH

_THRESHOLDS = {"easy": EASY_THRESH, "medium": MED_THRESH, "hard": HARD_THRESH}


# ──────────────────────────────────────────────────────────────────────────────
# Core evaluation function
# ──────────────────────────────────────────────────────────────────────────────

def evaluate_agent_on_task(
    agent,
    task_id: str,
    num_episodes: int = 10,
    seed_start:   int = 0,
    verbose:      bool = False,
) -> Dict[str, Any]:
    """
    Evaluate any agent on a specific task using the official task graders.

    Args:
        agent:        Agent with .act(observation) β†’ Action method.
        task_id:      "easy", "medium", or "hard".
        num_episodes: Episodes to run.
        seed_start:   Starting random seed.
        verbose:      Print per-episode stats.

    Returns:
        Dict with mean_score, std_score, success_rate, episode_scores, …
    """
    env          = AdaptiveAlertTriageEnv(task_id=task_id)
    is_hard      = task_id == "hard"
    threshold    = _THRESHOLDS[task_id]

    episode_scores   = []
    episode_rewards  = []
    episode_lengths  = []
    episode_failures = []

    for ep in range(num_episodes):
        obs = env.reset(seed=seed_start + ep)

        # ── Grader init (BUG FIX: HardTaskGrader takes NO args) ──────
        if task_id == "medium":
            grader = MediumTaskGrader(max_investigations_per_step=3)
        elif task_id == "hard":
            grader = HardTaskGrader()          # was wrongly HardTaskGrader(correlation_chains=[])
        else:
            grader = EasyTaskGrader()

        if hasattr(agent, "reset"):
            agent.reset()

        done         = False
        total_reward = 0.0
        steps        = 0

        while not done:
            if not obs.alerts:
                break

            try:
                action = agent.act(obs)
            except Exception as exc:
                if verbose:
                    print(f"  Agent error at step {steps}: {exc}")
                break

            next_obs, reward, done, info = env.step(action)

            # ── Hard task: update correlation state FIRST ─────────────
            if is_hard:
                grader.update_correlation_state(info.get("correlation_groups", []))

            # ── Grade every processed alert (BUG FIX: iterate all) ───
            for alert_data in info.get("processed_alerts", []):
                grader.process_step(alert_data, info)

            # ── Record failures (BUG FIX: correct method name + sig) ─
            if is_hard:
                grader.record_failures(info.get("failures_this_step", 0))   # was record_system_failure()

            total_reward += reward.value
            steps        += 1
            obs           = next_obs

        score = grader.get_episode_score()
        episode_scores.append(score)
        episode_rewards.append(total_reward)
        episode_lengths.append(steps)
        episode_failures.append(env.failures_count)

        if verbose:
            print(
                f"  ep {ep+1:3d}  score={score:.3f}  "
                f"reward={total_reward:+7.1f}  "
                f"steps={steps}  failures={env.failures_count}"
            )

    arr = np.array(episode_scores)
    return {
        "task_id":        task_id,
        "num_episodes":   num_episodes,
        "mean_score":     float(arr.mean()),
        "std_score":      float(arr.std()),
        "min_score":      float(arr.min()),
        "max_score":      float(arr.max()),
        "success_rate":   float((arr >= threshold).mean()),
        "mean_reward":    float(np.mean(episode_rewards)),
        "std_reward":     float(np.std(episode_rewards)),
        "mean_length":    float(np.mean(episode_lengths)),
        "mean_failures":  float(np.mean(episode_failures)),
        "episode_scores":   episode_scores,
        "episode_rewards":  episode_rewards,
    }


# ──────────────────────────────────────────────────────────────────────────────
# Full multi-agent evaluation
# ──────────────────────────────────────────────────────────────────────────────

def run_full_evaluation(
    num_episodes: int = 10,
    verbose:      bool = False,
) -> Dict[str, Dict[str, Any]]:
    """Run all baseline agents on all tasks."""
    agents = {
        "RuleBased":              RuleBasedAgent(),
        "ImprovedRuleBased":      ImprovedRuleBasedAgent(),
        "RuleBased_ResourceAware": RuleBasedAgent(resource_aware=True),
    }
    all_results: Dict[str, Dict[str, Any]] = {}

    for agent_name, agent in agents.items():
        if verbose:
            print(f"\n{'='*60}\nEvaluating: {agent_name}\n{'='*60}")
        agent_results: Dict[str, Any] = {}

        for task_id in ("easy", "medium", "hard"):
            if verbose:
                print(f"\n--- Task: {task_id} ---")
            res = evaluate_agent_on_task(
                agent=agent,
                task_id=task_id,
                num_episodes=num_episodes,
                verbose=verbose,
            )
            agent_results[task_id] = res
            if verbose:
                print(f"  mean={res['mean_score']:.3f}  "
                      f"success={res['success_rate']:.0%}  "
                      f"reward={res['mean_reward']:.1f}")

        all_results[agent_name] = agent_results

    return all_results


# ──────────────────────────────────────────────────────────────────────────────
# Display helpers
# ──────────────────────────────────────────────────────────────────────────────

def print_summary_table(all_results: Dict[str, Dict[str, Any]]) -> None:
    print("\n" + "=" * 80)
    print("EVALUATION SUMMARY")
    print("=" * 80)
    header = f"{'Agent':<28} {'Task':<10} {'MeanΒ±Std':>14}  {'Pass%':>8}  {'Failures':>9}"
    print(header)
    print("-" * 80)
    for agent_name, agent_results in all_results.items():
        for task_id, res in agent_results.items():
            print(
                f"{(agent_name if task_id == 'easy' else ''):<28} "
                f"{task_id:<10} "
                f"{res['mean_score']:.3f}Β±{res['std_score']:.3f}  "
                f"{res['success_rate']:>8.0%}  "
                f"{res['mean_failures']:>9.2f}"
            )
        print()


def save_results(
    all_results: Dict[str, Dict[str, Any]],
    filename: str = "evaluation_results.json",
) -> None:
    def _cvt(obj):
        if isinstance(obj, np.ndarray):     return obj.tolist()
        if isinstance(obj, (np.int64, np.int32)):   return int(obj)
        if isinstance(obj, (np.float64, np.float32)): return float(obj)
        if isinstance(obj, dict):  return {k: _cvt(v) for k, v in obj.items()}
        if isinstance(obj, list):  return [_cvt(v) for v in obj]
        return obj

    with open(filename, "w") as f:
        json.dump(_cvt(all_results), f, indent=2)
    print(f"\nResults saved β†’ {filename}")


# ──────────────────────────────────────────────────────────────────────────────
# CLI
# ──────────────────────────────────────────────────────────────────────────────

def main():
    parser = argparse.ArgumentParser(
        description="Evaluate baseline agents on Adaptive Alert Triage"
    )
    parser.add_argument("--episodes", type=int, default=10)
    parser.add_argument("--task",     choices=["easy", "medium", "hard", "all"], default="all")
    parser.add_argument("--agent",    choices=["rule", "improved", "resource", "all"], default="all")
    parser.add_argument("--verbose",  action="store_true")
    parser.add_argument("--output",   default="evaluation_results.json")
    args = parser.parse_args()

    print(f"Adaptive Alert Triage β€” Baseline Evaluation")
    print(f"Episodes/task: {args.episodes}  Task: {args.task}  Agent: {args.agent}\n")

    if args.task == "all" and args.agent == "all":
        all_results = run_full_evaluation(num_episodes=args.episodes, verbose=args.verbose)
    else:
        agents_map = {
            "rule":     ("RuleBased",               RuleBasedAgent()),
            "improved": ("ImprovedRuleBased",        ImprovedRuleBasedAgent()),
            "resource": ("RuleBased_ResourceAware",  RuleBasedAgent(resource_aware=True)),
        }
        agents = (
            {n: a for _, (n, a) in agents_map.items()}
            if args.agent == "all"
            else {agents_map[args.agent][0]: agents_map[args.agent][1]}
        )
        tasks  = ("easy", "medium", "hard") if args.task == "all" else (args.task,)
        all_results = {}
        for agent_name, agent in agents.items():
            agent_results = {}
            for task_id in tasks:
                agent_results[task_id] = evaluate_agent_on_task(
                    agent=agent, task_id=task_id,
                    num_episodes=args.episodes, verbose=args.verbose,
                )
            all_results[agent_name] = agent_results

    print_summary_table(all_results)
    if args.output:
        save_results(all_results, args.output)
    print("\nβœ… Evaluation complete!")


if __name__ == "__main__":
    main()