HarriziSaad commited on
Commit
8b239ea
Β·
verified Β·
1 Parent(s): 208b517

Update src/active_learning/al_loop.py

Browse files
Files changed (1) hide show
  1. src/active_learning/al_loop.py +3 -29
src/active_learning/al_loop.py CHANGED
@@ -1,19 +1,3 @@
1
- """
2
- al_loop.py β€” Pool-based active learning for transporter–compound discovery.
3
-
4
- Strategies implemented
5
- ----------------------
6
- random : baseline random acquisition
7
- uncertainty : select pairs with highest predictive entropy
8
- diversity : select pairs maximally different from already-labeled set
9
- causal : bias acquisition toward causally-ranked transporters
10
- hybrid : 0.5 * uncertainty + 0.5 * causal weight
11
-
12
- Usage
13
- -----
14
- python scripts/run_pipeline.py --task al --cfg env/config.yaml
15
- """
16
-
17
  from pathlib import Path
18
 
19
  import numpy as np
@@ -28,7 +12,6 @@ from ..atlas.model_mlp import AtlasMLP
28
  from ..utils.io import load_cfg, set_seed, save_json
29
 
30
 
31
- # ── Internal training helper ──────────────────────────────────────────────────
32
 
33
  def _train_model(ds: PairDataset, lr: float, epochs: int, batch_size: int,
34
  device: str) -> AtlasMLP:
@@ -53,8 +36,6 @@ def _predict_proba(model: AtlasMLP, ds: PairDataset, batch_size: int,
53
  return np.concatenate(probs)
54
 
55
 
56
- # ── Acquisition strategies ────────────────────────────────────────────────────
57
-
58
  def _scores_uncertainty(model, pool_ds, batch_size, device):
59
  """Predictive entropy: max at p=0.5."""
60
  p = _predict_proba(model, pool_ds, batch_size, device)
@@ -76,11 +57,10 @@ def _scores_diversity(model, pool_ds, labeled_ds, batch_size, device):
76
  pool_emb = _embeddings(pool_ds)
77
  labeled_emb = _embeddings(labeled_ds)
78
 
79
- # Cosine distance to nearest labeled point
80
  pool_n = pool_emb / (np.linalg.norm(pool_emb, axis=1, keepdims=True) + 1e-9)
81
  labeled_n = labeled_emb / (np.linalg.norm(labeled_emb, axis=1, keepdims=True) + 1e-9)
82
- sims = pool_n @ labeled_n.T # (n_pool, n_labeled)
83
- return 1.0 - sims.max(axis=1) # higher = more diverse
84
 
85
 
86
  def _scores_causal(pool_ds, causal_effects: dict) -> np.ndarray:
@@ -95,11 +75,10 @@ def _scores_causal(pool_ds, causal_effects: dict) -> np.ndarray:
95
  return weights / (weights.max() + 1e-9)
96
 
97
 
98
- # ── Main AL loop ──────────────────────────────────────────────────────────────
99
 
100
  def run_active_learning(
101
  cfg_path: str = "env/config.yaml",
102
- strategy: str = "uncertainty", # random | uncertainty | diversity | causal | hybrid
103
  causal_csv: str = "results/causal_effects.csv",
104
  ) -> dict:
105
  """
@@ -121,13 +100,11 @@ def run_active_learning(
121
  n = len(full_ds.pairs)
122
  rng = np.random.default_rng(tr_cfg["seed"])
123
 
124
- # Load causal weights if needed
125
  causal_effects = {}
126
  if strategy in ("causal", "hybrid") and Path(causal_csv).exists():
127
  df_c = pd.read_csv(causal_csv)
128
  causal_effects = dict(zip(df_c["gene"], df_c["ATE"].abs()))
129
 
130
- # Warm start
131
  init_k = int(al_cfg["init_frac"] * n)
132
  acquire_k = int(al_cfg["acquire_per_iter"] * n)
133
  labeled = set(rng.choice(n, size=init_k, replace=False).tolist())
@@ -145,7 +122,6 @@ def run_active_learning(
145
  model = _train_model(ds_labeled, tr_cfg["lr"], epochs=8,
146
  batch_size=tr_cfg["batch_size"], device=device)
147
 
148
- # ── Score pool ────────────────────────────────────────────────────────
149
  if strategy == "random":
150
  scores = rng.random(len(pool_list))
151
  elif strategy == "uncertainty":
@@ -161,14 +137,12 @@ def run_active_learning(
161
  else:
162
  raise ValueError(f"Unknown strategy: {strategy!r}")
163
 
164
- # ── Acquire top-k ─────────────────────────────────────────────────────
165
  acquire_k_actual = min(acquire_k, len(pool_list))
166
  top_local = np.argsort(scores)[::-1][:acquire_k_actual]
167
  newly_labeled = {pool_list[i] for i in top_local}
168
  labeled |= newly_labeled
169
  pool -= newly_labeled
170
 
171
- # ── Evaluate on held-out pool ─────────────────────────────────────────
172
  hold_size = min(int(0.2 * n), len(pool))
173
  if hold_size > 0:
174
  hold_idx = rng.choice(sorted(pool), size=hold_size, replace=False)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  from pathlib import Path
2
 
3
  import numpy as np
 
12
  from ..utils.io import load_cfg, set_seed, save_json
13
 
14
 
 
15
 
16
  def _train_model(ds: PairDataset, lr: float, epochs: int, batch_size: int,
17
  device: str) -> AtlasMLP:
 
36
  return np.concatenate(probs)
37
 
38
 
 
 
39
  def _scores_uncertainty(model, pool_ds, batch_size, device):
40
  """Predictive entropy: max at p=0.5."""
41
  p = _predict_proba(model, pool_ds, batch_size, device)
 
57
  pool_emb = _embeddings(pool_ds)
58
  labeled_emb = _embeddings(labeled_ds)
59
 
 
60
  pool_n = pool_emb / (np.linalg.norm(pool_emb, axis=1, keepdims=True) + 1e-9)
61
  labeled_n = labeled_emb / (np.linalg.norm(labeled_emb, axis=1, keepdims=True) + 1e-9)
62
+ sims = pool_n @ labeled_n.T
63
+ return 1.0 - sims.max(axis=1)
64
 
65
 
66
  def _scores_causal(pool_ds, causal_effects: dict) -> np.ndarray:
 
75
  return weights / (weights.max() + 1e-9)
76
 
77
 
 
78
 
79
  def run_active_learning(
80
  cfg_path: str = "env/config.yaml",
81
+ strategy: str = "uncertainty",
82
  causal_csv: str = "results/causal_effects.csv",
83
  ) -> dict:
84
  """
 
100
  n = len(full_ds.pairs)
101
  rng = np.random.default_rng(tr_cfg["seed"])
102
 
 
103
  causal_effects = {}
104
  if strategy in ("causal", "hybrid") and Path(causal_csv).exists():
105
  df_c = pd.read_csv(causal_csv)
106
  causal_effects = dict(zip(df_c["gene"], df_c["ATE"].abs()))
107
 
 
108
  init_k = int(al_cfg["init_frac"] * n)
109
  acquire_k = int(al_cfg["acquire_per_iter"] * n)
110
  labeled = set(rng.choice(n, size=init_k, replace=False).tolist())
 
122
  model = _train_model(ds_labeled, tr_cfg["lr"], epochs=8,
123
  batch_size=tr_cfg["batch_size"], device=device)
124
 
 
125
  if strategy == "random":
126
  scores = rng.random(len(pool_list))
127
  elif strategy == "uncertainty":
 
137
  else:
138
  raise ValueError(f"Unknown strategy: {strategy!r}")
139
 
 
140
  acquire_k_actual = min(acquire_k, len(pool_list))
141
  top_local = np.argsort(scores)[::-1][:acquire_k_actual]
142
  newly_labeled = {pool_list[i] for i in top_local}
143
  labeled |= newly_labeled
144
  pool -= newly_labeled
145
 
 
146
  hold_size = min(int(0.2 * n), len(pool))
147
  if hold_size > 0:
148
  hold_idx = rng.choice(sorted(pool), size=hold_size, replace=False)