""" EXP Q ADDENDUM: add multi-property discrete + continuous configs to the scatter sweep so the headline-PosDis configs (0.76 discrete, 0.40 continuous multi-prop) are represented in the metric-vs-transfer correlation. Trains: disc_multi_L3_V5 — discrete bottleneck, 2 head per agent, V=5, multi-prop (mass_bin + restit_bin, 2-headed receiver) disc_multi_L4_V10 — same but L=4, V=10 (broader coverage) cont_multi_dim3 — continuous bottleneck, code_dim=3 per agent, multi-prop Per config: - Within: 3 seeds, mean across 2 heads (mass + restit) for combined acc - Metrics: TopSim, PosDis, CausalSpec on multi-prop labels - Cross to ramp at N=16 and N=192 (restitution only, since ramp lacks mass) Re-uses EXP N's MultiPropReceiver and the metric helpers in EXP N + Q. """ import json, time, sys, os, math from pathlib import Path from datetime import datetime, timezone import numpy as np import torch import torch.nn as nn import torch.nn.functional as F sys.path.insert(0, os.path.dirname(__file__)) from _kinematics_train import ( DEVICE, ClassifierReceiver, HIDDEN_DIM, N_AGENTS, BATCH_SIZE, SENDER_LR, RECEIVER_LR, EARLY_STOP_PATIENCE, ) from _killer_experiment import TemporalEncoder, DiscreteSender, DiscreteMultiSender from _overnight_p1_transfer import make_splits from _overnight_p3_matrix import load_labels, load_feat_subsampled from _rev_f_cnn_control import ci95 from _rev_q_posdis_scatter import build_discrete_sender, discrete_token_extract, discrete_topsim from _rev_n_multiprop_continuous import ( MultiPropReceiver, train_multiprop_continuous_base, topsim_multiprop, posdis_multiprop, causal_spec_multiprop, ) from _rev_m_continuous_bottleneck import ( build_continuous_sender, get_continuous_messages, train_recv_frozen_cont, ) OUT = Path("results/reviewer_response/exp_q_addendum") OUT.mkdir(parents=True, exist_ok=True) N_SEEDS = 3 N_LIST = [16, 192] def log(msg): ts = datetime.now(timezone.utc).strftime("%H:%M:%SZ") print(f"[{ts}] EXP-QADD: {msg}", flush=True) # ─── Discrete multi-prop training ─── def train_discrete_multi(feat, labels_list, seed, n_heads, vocab_size, n_epochs=150): """Train DiscreteSender with multi-prop receiver (2 heads per receiver).""" N, nf, dim = feat.shape fpa = 1 msg_dim = vocab_size * n_heads * N_AGENTS agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)] torch.manual_seed(seed); np.random.seed(seed) rng = np.random.RandomState(seed * 1000 + 42) primary = labels_list[0] train_ids, holdout_ids = [], [] for c in np.unique(primary): ids_c = np.where(primary == c)[0] rng.shuffle(ids_c) split = max(1, len(ids_c) // 5) holdout_ids.extend(ids_c[:split]); train_ids.extend(ids_c[split:]) train_ids = np.array(train_ids); holdout_ids = np.array(holdout_ids) n_classes_per_prop = [int(lbl.max()) + 1 for lbl in labels_list] chance = 1.0 / max(n_classes_per_prop) sender = build_discrete_sender(dim, n_heads, vocab_size, fpa) receivers = [MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE) for _ in range(3)] so = torch.optim.Adam(sender.parameters(), lr=SENDER_LR) ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers] labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list] me = math.log(vocab_size) n_batches = max(1, len(train_ids) // BATCH_SIZE) best_acc = 0.0; best_ep = 0 best_sender_state = None; best_receiver_states = None; best_recv_idx = 0 for ep in range(n_epochs): if ep - best_ep > EARLY_STOP_PATIENCE and best_acc > chance + 0.05: break if ep > 0 and ep % 40 == 0: for i in range(len(receivers)): receivers[i] = MultiPropReceiver(msg_dim, HIDDEN_DIM, n_classes_per_prop).to(DEVICE) ros[i] = torch.optim.Adam(receivers[i].parameters(), lr=RECEIVER_LR) sender.train(); [r.train() for r in receivers] tau = 3.0 + (1.0 - 3.0) * ep / max(1, n_epochs - 1) hard = ep >= 30 rng_ep = np.random.RandomState(seed * 10000 + ep) perm = rng_ep.permutation(train_ids) for b in range(n_batches): batch_ids = perm[b*BATCH_SIZE:(b+1)*BATCH_SIZE] if len(batch_ids) < 4: continue views = [v[batch_ids].to(DEVICE) for v in agent_views] tgts = [ld[batch_ids] for ld in labels_dev] msg, logits_list = sender(views, tau=tau, hard=hard) loss = torch.tensor(0.0, device=DEVICE) for r in receivers: head_logits = r(msg) for hl, tgt in zip(head_logits, tgts): loss = loss + F.cross_entropy(hl, tgt) loss = loss / (len(receivers) * len(tgts)) for lg in logits_list: lp = F.log_softmax(lg, -1); p = lp.exp().clamp(min=1e-8) ent = -(p * lp).sum(-1).mean() if ent / me < 0.1: loss = loss - 0.03 * ent if torch.isnan(loss): so.zero_grad(); [o.zero_grad() for o in ros]; continue so.zero_grad(); [o.zero_grad() for o in ros] loss.backward() torch.nn.utils.clip_grad_norm_(sender.parameters(), 1.0) so.step(); [o.step() for o in ros] if ep % 50 == 0 and DEVICE.type == "mps": torch.mps.empty_cache() if (ep + 1) % 10 == 0 or ep == 0: sender.eval(); [r.eval() for r in receivers] with torch.no_grad(): v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] msg_ho, _ = sender(v_ho) tgt_ho = [ld[holdout_ids] for ld in labels_dev] best_per_recv = 0.0; best_idx = 0 for ri, r in enumerate(receivers): head_logits = r(msg_ho) accs = [(hl.argmax(-1) == tgt).float().mean().item() for hl, tgt in zip(head_logits, tgt_ho)] combined = float(np.mean(accs)) if combined > best_per_recv: best_per_recv = combined; best_idx = ri if best_per_recv > best_acc: best_acc = best_per_recv; best_ep = ep best_sender_state = {k: v.cpu().clone() for k, v in sender.state_dict().items()} best_receiver_states = [ {k: v.cpu().clone() for k, v in r.state_dict().items()} for r in receivers] best_recv_idx = best_idx return { "sender_state": best_sender_state, "receiver_states": best_receiver_states, "best_recv_idx": best_recv_idx, "train_ids": train_ids, "holdout_ids": holdout_ids, "task_acc": best_acc, "chance": chance, "n_classes_per_prop": n_classes_per_prop, "fpa": 1, "dim": dim, "n_heads": n_heads, "vocab_size": vocab_size, "msg_dim": msg_dim, } # ─── Discrete multi-prop metrics ─── def discrete_multi_topsim(tokens, labels_list, n_pairs=5000): """Spearman corr between Hamming(message tokens) and L1(label vector).""" from scipy.stats import spearmanr rng = np.random.RandomState(42) N = tokens.shape[0] n_pairs = min(n_pairs, N * (N - 1) // 2) tok_d = []; lbl_d = [] seen = set() for _ in range(n_pairs): i, j = rng.randint(0, N), rng.randint(0, N) if i == j or (i, j) in seen or (j, i) in seen: continue seen.add((i, j)) tok_d.append(int((tokens[i] != tokens[j]).sum())) lbl_d.append(sum(abs(int(lbl[i]) - int(lbl[j])) for lbl in labels_list)) if len(tok_d) < 10 or np.std(tok_d) < 1e-9 or np.std(lbl_d) < 1e-9: return float("nan") rho, _ = spearmanr(tok_d, lbl_d) return float(rho) if not np.isnan(rho) else 0.0 def _mi_disc(x, y): n = len(x) n_x = int(np.max(x)) + 1; n_y = int(np.max(y)) + 1 p_x = np.bincount(x, minlength=n_x) / n p_y = np.bincount(y, minlength=n_y) / n H_x = -np.sum([p * np.log(p) for p in p_x if p > 0]) H_y = -np.sum([p * np.log(p) for p in p_y if p > 0]) joint = np.zeros((n_x, n_y)) for xv, yv in zip(x, y): joint[int(xv), int(yv)] += 1 joint /= n H_xy = 0.0 for v in joint.ravel(): if v > 0: H_xy -= v * np.log(v) return max(H_x + H_y - H_xy, 0.0) def discrete_multi_posdis(tokens, labels_list): """Standard PosDis on discrete tokens: per-position, MI with each property; PosDis = mean over positions of (top-second)/top.""" P = tokens.shape[1] K = len(labels_list) mi_matrix = np.zeros((P, K)) for p in range(P): for k in range(K): mi_matrix[p, k] = _mi_disc(tokens[:, p], labels_list[k]) if mi_matrix.sum() < 1e-9: return float("nan"), mi_matrix n_active = 0; total = 0.0 for p in range(P): sorted_mi = np.sort(mi_matrix[p])[::-1] if sorted_mi[0] > 1e-6: total += (sorted_mi[0] - sorted_mi[1]) / sorted_mi[0] n_active += 1 if n_active == 0: return float("nan"), mi_matrix return float(total / n_active), mi_matrix def discrete_multi_causal(base, feat, labels_list, holdout_ids): """Mask each (agent x head) block; per-property accuracy drop.""" sender = build_discrete_sender(feat.shape[2], base["n_heads"], base["vocab_size"], base["fpa"]) sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE) receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes_per_prop"]).to(DEVICE) for _ in range(len(base["receiver_states"]))] for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s) [r.eval() for r in receivers] best_recv = receivers[base.get("best_recv_idx", 0)] agent_views = [feat[:, i:i+1, :] for i in range(N_AGENTS)] labels_dev = [torch.tensor(lbl, dtype=torch.long).to(DEVICE) for lbl in labels_list] K = len(labels_list); V = base["vocab_size"]; H = base["n_heads"] n_positions = N_AGENTS * H with torch.no_grad(): v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] msg_ho, _ = sender(v_ho) tgt_ho = [ld[holdout_ids] for ld in labels_dev] baseline_per_prop = [(hl.argmax(-1) == tgt).float().mean().item() for hl, tgt in zip(best_recv(msg_ho), tgt_ho)] drops = np.zeros((n_positions, K)) for pos in range(n_positions): masked = msg_ho.clone() start = pos * V; end = start + V mean_block = msg_ho[:, start:end].mean(dim=0) masked[:, start:end] = mean_block for p_idx, (hl, tgt) in enumerate(zip(best_recv(masked), tgt_ho)): acc = (hl.argmax(-1) == tgt).float().mean().item() drops[pos, p_idx] = baseline_per_prop[p_idx] - acc return baseline_per_prop, drops # ─── Cross-scenario eval (single-property restitution on ramp) ─── def disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, ho_ids, restit_idx=1): """Apply discrete-multi sender + best receiver's restit head to target.""" sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"], base["vocab_size"], base["fpa"]) sender.load_state_dict(base["sender_state"]); sender.eval().to(DEVICE) receivers = [MultiPropReceiver(base["msg_dim"], HIDDEN_DIM, base["n_classes_per_prop"]).to(DEVICE) for _ in range(len(base["receiver_states"]))] for r, s in zip(receivers, base["receiver_states"]): r.load_state_dict(s) [r.eval() for r in receivers] agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)] labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE) with torch.no_grad(): v_ho = [v[ho_ids].to(DEVICE) for v in agent_views] msg_ho, _ = sender(v_ho) tgt_ho = labels_dev[ho_ids] best = 0.0 for r in receivers: head_logits = r(msg_ho) preds = head_logits[restit_idx].argmax(-1) acc = (preds == tgt_ho).float().mean().item() if acc > best: best = acc return best def disc_multi_train_recv_frozen(base, feat_tgt, labels_tgt, train_ids, holdout_ids, seed, n_target, n_epochs=80): """Freeze sender; train fresh single-property receiver on n_target stratified target examples; eval on holdout. Mirrors disc_train_recv_custom but uses multi-prop sender.""" if n_target == 0: return disc_multi_zero_shot_restit(base, feat_tgt, labels_tgt, holdout_ids) rng = np.random.RandomState(seed * 311 + 7 + n_target) n_t_classes = int(np.max(labels_tgt)) + 1 per_class = max(1, n_target // n_t_classes) picks = [] for c in range(n_t_classes): ids_c = np.array([i for i in train_ids if labels_tgt[i] == c]) if len(ids_c) == 0: continue rng.shuffle(ids_c) picks.extend(ids_c[:per_class]) picks = np.array(picks) if len(picks) > n_target: picks = picks[:n_target] elif len(picks) < n_target and len(train_ids) > len(picks): extras = np.array([i for i in train_ids if i not in set(picks)]) rng.shuffle(extras) picks = np.concatenate([picks, extras[:n_target - len(picks)]]) if len(picks) < 2: return float("nan") sender = build_discrete_sender(feat_tgt.shape[2], base["n_heads"], base["vocab_size"], base["fpa"]) sender.load_state_dict(base["sender_state"]); sender.to(DEVICE).eval() for p in sender.parameters(): p.requires_grad = False receivers = [ClassifierReceiver(base["msg_dim"], HIDDEN_DIM, n_t_classes).to(DEVICE) for _ in range(3)] ros = [torch.optim.Adam(r.parameters(), lr=RECEIVER_LR) for r in receivers] agent_views = [feat_tgt[:, i:i+1, :] for i in range(N_AGENTS)] labels_dev = torch.tensor(labels_tgt, dtype=torch.long).to(DEVICE) bs = min(BATCH_SIZE, len(picks)) best = 0.0 for ep in range(n_epochs): [r.train() for r in receivers] rng_ep = np.random.RandomState(seed * 10000 + ep) perm = rng_ep.permutation(picks) for b in range(max(1, len(picks) // bs)): batch = perm[b*bs:(b+1)*bs] if len(batch) < 2: continue views = [v[batch].to(DEVICE) for v in agent_views] with torch.no_grad(): msg, _ = sender(views) for r, o in zip(receivers, ros): logits = r(msg) loss = F.cross_entropy(logits, labels_dev[batch]) if torch.isnan(loss): continue o.zero_grad(); loss.backward(); o.step() if (ep + 1) % 5 == 0: [r.eval() for r in receivers] with torch.no_grad(): v_ho = [v[holdout_ids].to(DEVICE) for v in agent_views] msg_ho, _ = sender(v_ho) tgt_ho = labels_dev[holdout_ids] for r in receivers: preds = r(msg_ho).argmax(-1) acc = (preds == tgt_ho).float().mean().item() if acc > best: best = acc return best # ─── Main ─── def main(): t0 = time.time() log("=" * 60) log("EXP Q ADDENDUM: multi-property bottleneck configs") feat_c = load_feat_subsampled("collision", "vjepa2") feat_r = load_feat_subsampled("ramp", "vjepa2") lbl_c_mass = load_labels("collision", "mass") lbl_c_rest = load_labels("collision", "restitution") lbl_r_rest = load_labels("ramp", "restitution") log(f" collision: feat={tuple(feat_c.shape)} mass={np.bincount(lbl_c_mass).tolist()} " f"rest={np.bincount(lbl_c_rest).tolist()}") rows = [] # ── Discrete multi-prop configs ── discrete_specs = [ ("disc_multi_L3_V5", 3, 5), ("disc_multi_L4_V10", 4, 10), ] for name, H, V in discrete_specs: log(f"\n --- {name} (L={H}, V={V}, multi-prop) ---") within_accs = []; bases = [] for seed in range(N_SEEDS): t_s = time.time() try: base = train_discrete_multi(feat_c, [lbl_c_mass, lbl_c_rest], seed, H, V) bases.append(base); within_accs.append(float(base["task_acc"])) log(f" {name} s{seed}: combined within={base['task_acc']:.3f} " f"[{time.time()-t_s:.0f}s]") except Exception as e: log(f" {name} s{seed} FAILED: {e}") bases.append(None); within_accs.append(float("nan")) valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)] if not valid: log(f" {name}: no successful base"); continue best_idx = max(valid, key=lambda x: x[1])[0] best_base = bases[best_idx] ho_ids = best_base["holdout_ids"] # Within metrics on best try: tokens = discrete_token_extract(best_base, feat_c) tokens_ho = tokens[ho_ids] ts = discrete_multi_topsim(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) pd_, mi = discrete_multi_posdis(tokens_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) base_pp, drops = discrete_multi_causal(best_base, feat_c, [lbl_c_mass, lbl_c_rest], ho_ids) cs = float(drops.max()) except Exception as e: log(f" {name} metrics FAILED: {e}") ts = pd_ = cs = float("nan") # Cross-scenario coll->ramp at N=16, N=192 (restitution) cross = {n: [] for n in N_LIST} for seed, base in enumerate(bases): if base is None: for n in N_LIST: cross[n].append(float("nan")) continue tr_t, ho_t = make_splits(lbl_r_rest, seed) for n in N_LIST: try: acc = disc_multi_train_recv_frozen(base, feat_r, lbl_r_rest, tr_t, ho_t, seed, n) cross[n].append(float(acc)) except Exception as e: log(f" {name} s{seed} N={n} FAILED: {e}") cross[n].append(float("nan")) wm = float(np.mean([a for a in within_accs if not np.isnan(a)])) cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)])) if any(not np.isnan(x) for x in cross[n]) else float("nan") for n in N_LIST} log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} " f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}") rows.append({ "name": name, "type": "discrete_multi", "n_heads": H, "vocab_size": V, "msg_dim": V * H * N_AGENTS, "within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs, "cross_n16": cm[16], "cross_n192": cm[192], }) # ── Continuous multi-prop config (matches Exp N exactly) ── cont_spec = ("cont_multi_dim3", 3) name, D = cont_spec log(f"\n --- {name} (D={D}, multi-prop) ---") within_accs = []; bases = [] for seed in range(N_SEEDS): t_s = time.time() try: base = train_multiprop_continuous_base( feat_c, [lbl_c_mass, lbl_c_rest], seed, code_dim_per_agent=D, n_epochs=150) bases.append(base); within_accs.append(float(base["task_acc"])) log(f" {name} s{seed}: combined within={base['task_acc']:.3f} [{time.time()-t_s:.0f}s]") except Exception as e: log(f" {name} s{seed} FAILED: {e}") bases.append(None); within_accs.append(float("nan")) valid = [(i, a) for i, a in enumerate(within_accs) if not np.isnan(a)] if valid: best_idx = max(valid, key=lambda x: x[1])[0] best_base = bases[best_idx] ho_ids = best_base["holdout_ids"] try: msgs = get_continuous_messages(best_base, feat_c) msgs_ho = msgs[ho_ids] ts = topsim_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) pd_, _ = posdis_multiprop(msgs_ho, [lbl_c_mass[ho_ids], lbl_c_rest[ho_ids]]) base_pp, drops = causal_spec_multiprop(best_base, feat_c, [lbl_c_mass, lbl_c_rest], ho_ids) cs = float(drops.max()) except Exception as e: log(f" {name} metrics FAILED: {e}") ts = pd_ = cs = float("nan") # Cross to ramp on restitution: build single-task base view; train fresh receiver cross = {n: [] for n in N_LIST} for seed, base in enumerate(bases): if base is None: for n in N_LIST: cross[n].append(float("nan")) continue single_base = dict(base) single_base["n_classes"] = base["n_classes_per_prop"][1] single_base["receiver_states"] = [] tr_t, ho_t = make_splits(lbl_r_rest, seed) for n in N_LIST: try: acc = train_recv_frozen_cont(single_base, feat_r, lbl_r_rest, tr_t, ho_t, seed, n) cross[n].append(float(acc)) except Exception as e: log(f" {name} s{seed} N={n} FAILED: {e}") cross[n].append(float("nan")) wm = float(np.mean([a for a in within_accs if not np.isnan(a)])) cm = {n: float(np.mean([x for x in cross[n] if not np.isnan(x)])) if any(not np.isnan(x) for x in cross[n]) else float("nan") for n in N_LIST} log(f" {name}: within={wm:.3f} TopSim={ts:.3f} PosDis={pd_:.3f} " f"CausalSpec={cs:.3f} cross16={cm[16]:.3f} cross192={cm[192]:.3f}") rows.append({ "name": name, "type": "continuous_multi", "code_dim": D, "msg_dim": D * N_AGENTS, "within": wm, "topsim": ts, "posdis": pd_, "causal_spec": cs, "cross_n16": cm[16], "cross_n192": cm[192], }) # ── Combine with original Exp Q rows for full correlation ── original_q_rows = [ # (name, type, topsim, posdis, causal_spec, cross16, cross192) ("disc_L2_V5", "discrete", 0.88, 0.20, 0.02, 41.7, 43.9), ("disc_L2_V10", "discrete", 0.84, 0.25, 0.05, 46.1, 41.7), ("disc_L3_V5", "discrete", 0.84, 0.13, 0.02, 43.3, 42.8), ("disc_L3_V10", "discrete", 0.84, 0.12, 0.01, 43.3, 45.6), ("disc_L4_V5", "discrete", 0.90, 0.10, 0.01, 41.1, 42.2), ("disc_L4_V10", "discrete", 0.82, 0.08, 0.02, 45.0, 45.0), ("disc_L5_V5", "discrete", 0.89, 0.07, 0.02, 40.0, 43.9), ("cont_dim2", "continuous", 0.92, 0.15, 0.20, 48.9, 54.4), ("cont_dim3", "continuous", 0.91, 0.15, 0.02, 40.6, 41.1), ("cont_dim5", "continuous", 0.89, 0.06, 0.03, 47.2, 43.9), ("cont_dim10", "continuous", 0.88, 0.04, 0.01, 47.8, 48.3), ("cont_dim20", "continuous", 0.90, 0.02, 0.00, 48.9, 55.0), ] all_rows = list(original_q_rows) for r in rows: all_rows.append(( r["name"], r["type"], r["topsim"], r["posdis"], r["causal_spec"], r["cross_n16"] * 100 if r["cross_n16"] <= 1 else r["cross_n16"], r["cross_n192"] * 100 if r["cross_n192"] <= 1 else r["cross_n192"], )) # Spearman across all rows from scipy.stats import spearmanr def safe_corr(idx_x, idx_y): x = []; y = [] for r in all_rows: if not (np.isnan(r[idx_x]) or np.isnan(r[idx_y])): x.append(r[idx_x]); y.append(r[idx_y]) if len(x) < 4 or np.std(x) < 1e-9 or np.std(y) < 1e-9: return float("nan"), float("nan") rho, p = spearmanr(x, y) return float(rho), float(p) corrs = { ("topsim", 16): safe_corr(2, 5), ("topsim", 192): safe_corr(2, 6), ("posdis", 16): safe_corr(3, 5), ("posdis", 192): safe_corr(3, 6), ("causal", 16): safe_corr(4, 5), ("causal", 192): safe_corr(4, 6), } # Print summary lines = [ "EXP Q ADDENDUM -- multi-property bottleneck configs", "", f"{'Config':<22s} | {'TopSim':<8s} | {'PosDis':<8s} | {'CausalSpec':<12s} | " f"{'Cross 16':<10s} | {'Cross 192':<10s}", "-" * 90, ] for r in all_rows: name, typ, ts, pd_, cs, c16, c192 = r lines.append(f"{name:<22s} | {ts:+.2f} | {pd_:.2f} | {cs:.2f} " f"| {c16:5.1f}% | {c192:5.1f}%") lines.append("") lines.append("FULL-SWEEP SPEARMAN (now including multi-prop configs):") for tgt_n in [16, 192]: lines.append(f" vs cross_n{tgt_n}:") for met, label in [("topsim", "TopSim"), ("posdis", "PosDis"), ("causal", "CausalSpec")]: rho, p = corrs[(met, tgt_n)] lines.append(f" {label:<12s}: rho={rho:+.2f} p={p:.3f}") abs_max_rho = 0 for k, (rho, p) in corrs.items(): if not np.isnan(rho): abs_max_rho = max(abs_max_rho, abs(rho)) lines.append("") if abs_max_rho < 0.30: verd = "Metrics still uncorrelated with transfer." elif abs_max_rho < 0.55: verd = f"Weak/moderate correlation (max |rho|={abs_max_rho:.2f})." else: verd = f"Strong correlation (max |rho|={abs_max_rho:.2f}). Reframe." lines.append(f"VERDICT: {verd}") lines.append(f"\nTotal runtime: {(time.time()-t0)/60:.1f} min") summary = "\n".join(lines) (OUT / "exp_q_addendum_summary.txt").write_text(summary + "\n") (OUT / "exp_q_addendum_summary.json").write_text(json.dumps({ "new_rows": rows, "all_rows_combined": [{"name": r[0], "type": r[1], "topsim": r[2], "posdis": r[3], "causal_spec": r[4], "cross_n16": r[5], "cross_n192": r[6]} for r in all_rows], "spearman": {f"{m}__n{n}": list(v) for (m, n), v in corrs.items()}, }, indent=2, default=str)) print("\n" + summary, flush=True) log(f"DONE in {(time.time()-t0)/60:.1f} min") if __name__ == "__main__": main()