fix(P0): Implement SPEC_06 - Simple Mode Synthesis Fix (#71)
Browse files* feat: implement SPEC_06 fixes for simple mode synthesis
Implements research-backed architecture changes for Simple mode:
- Diverse evidence selection with lost-in-the-middle mitigation
- Debiased judge prompt (scoring only, no decision making)
- Code-enforced termination criteria (prevents infinite loops)
- Real partial synthesis with drug candidates when max iterations reached
- Unit and integration tests for all new logic
* fix: address CodeRabbit review feedback
- Fix type annotation: recommendation str -> Literal["continue", "synthesize"]
- Add
@pytest
.mark.unit markers to unit tests
- Add
@pytest
.mark.integration markers to integration tests
- Simplify early/late evidence assertions using set intersection
- Replace next() with explicit filtering for clearer test diagnostics
- Extract hardcoded thresholds (50, 30) to TERMINATION_CRITERIA
- Fix e2e mock to accept iteration/max_iterations parameters
All 177 tests pass, mypy clean, ruff clean.
* docs: add P3 architectural gap report for ephemeral memory
- Introduced a new document outlining the architectural gap related to the use of ephemeral memory in the EmbeddingService.
- Highlights the lack of persistence in the current implementation, leading to no long-term learning and redundant costs.
- Provides technical details and recommendations for addressing the issue in future product iterations.
- docs/bugs/P3_ARCHITECTURAL_GAP_EPHEMERAL_MEMORY.md +23 -0
- src/agent_factory/judges.py +27 -2
- src/orchestrators/base.py +9 -1
- src/orchestrators/simple.py +177 -22
- src/prompts/judge.py +97 -27
- tests/e2e/conftest.py +1 -1
- tests/integration/test_simple_mode_synthesis.py +147 -0
- tests/unit/orchestrators/test_termination.py +104 -0
- tests/unit/prompts/test_judge_prompt.py +61 -0
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# P3: Ephemeral Memory Architecture (No Persistence)
|
| 2 |
+
|
| 3 |
+
**Status:** OPEN
|
| 4 |
+
**Priority:** P3 (Feature/Architecture Gap)
|
| 5 |
+
**Found By:** Codebase Investigation
|
| 6 |
+
**Date:** 2025-11-29
|
| 7 |
+
|
| 8 |
+
## Description
|
| 9 |
+
The current `EmbeddingService` (`src/services/embeddings.py`) initializes an **in-memory** ChromaDB client (`chromadb.Client()`) and creates a random UUID-based collection for every new session.
|
| 10 |
+
|
| 11 |
+
While `src/utils/config.py` defines a `chroma_db_path` for persistence, it is currently **ignored**.
|
| 12 |
+
|
| 13 |
+
## Impact
|
| 14 |
+
1. **No Long-Term Learning:** The agent cannot "remember" research from previous runs. Every time you restart the app, it starts from zero.
|
| 15 |
+
2. **Redundant Costs:** If a user researches "Diabetes" twice, the agent re-searches and re-embeds the same papers, wasting tokens and compute time.
|
| 16 |
+
|
| 17 |
+
## Technical Details
|
| 18 |
+
- **Current:** `self._client = chromadb.Client()` (In-Memory)
|
| 19 |
+
- **Required:** `self._client = chromadb.PersistentClient(path=settings.chroma_db_path)`
|
| 20 |
+
|
| 21 |
+
## Recommendation
|
| 22 |
+
For a "Hackathon Demo," this is **low priority** (ephemeral is fine).
|
| 23 |
+
For a "Real Product," this is **critical** (users expect a library of research).
|
|
@@ -19,6 +19,7 @@ from src.prompts.judge import (
|
|
| 19 |
SYSTEM_PROMPT,
|
| 20 |
format_empty_evidence_prompt,
|
| 21 |
format_user_prompt,
|
|
|
|
| 22 |
)
|
| 23 |
from src.utils.config import settings
|
| 24 |
from src.utils.models import AssessmentDetails, Evidence, JudgeAssessment
|
|
@@ -102,6 +103,8 @@ class JudgeHandler:
|
|
| 102 |
self,
|
| 103 |
question: str,
|
| 104 |
evidence: list[Evidence],
|
|
|
|
|
|
|
| 105 |
) -> JudgeAssessment:
|
| 106 |
"""
|
| 107 |
Assess evidence and determine if it's sufficient.
|
|
@@ -109,6 +112,8 @@ class JudgeHandler:
|
|
| 109 |
Args:
|
| 110 |
question: The user's research question
|
| 111 |
evidence: List of Evidence objects from search
|
|
|
|
|
|
|
| 112 |
|
| 113 |
Returns:
|
| 114 |
JudgeAssessment with evaluation results
|
|
@@ -120,11 +125,20 @@ class JudgeHandler:
|
|
| 120 |
"Starting evidence assessment",
|
| 121 |
question=question[:100],
|
| 122 |
evidence_count=len(evidence),
|
|
|
|
| 123 |
)
|
| 124 |
|
| 125 |
# Format the prompt based on whether we have evidence
|
| 126 |
if evidence:
|
| 127 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 128 |
else:
|
| 129 |
user_prompt = format_empty_evidence_prompt(question)
|
| 130 |
|
|
@@ -218,6 +232,8 @@ class HFInferenceJudgeHandler:
|
|
| 218 |
self,
|
| 219 |
question: str,
|
| 220 |
evidence: list[Evidence],
|
|
|
|
|
|
|
| 221 |
) -> JudgeAssessment:
|
| 222 |
"""
|
| 223 |
Assess evidence using HuggingFace Inference API.
|
|
@@ -246,7 +262,14 @@ class HFInferenceJudgeHandler:
|
|
| 246 |
|
| 247 |
# Format the user prompt
|
| 248 |
if evidence:
|
| 249 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 250 |
else:
|
| 251 |
user_prompt = format_empty_evidence_prompt(question)
|
| 252 |
|
|
@@ -535,6 +558,8 @@ class MockJudgeHandler:
|
|
| 535 |
self,
|
| 536 |
question: str,
|
| 537 |
evidence: list[Evidence],
|
|
|
|
|
|
|
| 538 |
) -> JudgeAssessment:
|
| 539 |
"""Return assessment based on actual evidence (demo mode)."""
|
| 540 |
self.call_count += 1
|
|
|
|
| 19 |
SYSTEM_PROMPT,
|
| 20 |
format_empty_evidence_prompt,
|
| 21 |
format_user_prompt,
|
| 22 |
+
select_evidence_for_judge,
|
| 23 |
)
|
| 24 |
from src.utils.config import settings
|
| 25 |
from src.utils.models import AssessmentDetails, Evidence, JudgeAssessment
|
|
|
|
| 103 |
self,
|
| 104 |
question: str,
|
| 105 |
evidence: list[Evidence],
|
| 106 |
+
iteration: int = 0,
|
| 107 |
+
max_iterations: int = 10,
|
| 108 |
) -> JudgeAssessment:
|
| 109 |
"""
|
| 110 |
Assess evidence and determine if it's sufficient.
|
|
|
|
| 112 |
Args:
|
| 113 |
question: The user's research question
|
| 114 |
evidence: List of Evidence objects from search
|
| 115 |
+
iteration: Current iteration number
|
| 116 |
+
max_iterations: Maximum allowed iterations
|
| 117 |
|
| 118 |
Returns:
|
| 119 |
JudgeAssessment with evaluation results
|
|
|
|
| 125 |
"Starting evidence assessment",
|
| 126 |
question=question[:100],
|
| 127 |
evidence_count=len(evidence),
|
| 128 |
+
iteration=iteration,
|
| 129 |
)
|
| 130 |
|
| 131 |
# Format the prompt based on whether we have evidence
|
| 132 |
if evidence:
|
| 133 |
+
# Select diverse evidence using embeddings (if available)
|
| 134 |
+
selected_evidence = await select_evidence_for_judge(evidence, question)
|
| 135 |
+
user_prompt = format_user_prompt(
|
| 136 |
+
question,
|
| 137 |
+
selected_evidence,
|
| 138 |
+
iteration,
|
| 139 |
+
max_iterations,
|
| 140 |
+
total_evidence_count=len(evidence),
|
| 141 |
+
)
|
| 142 |
else:
|
| 143 |
user_prompt = format_empty_evidence_prompt(question)
|
| 144 |
|
|
|
|
| 232 |
self,
|
| 233 |
question: str,
|
| 234 |
evidence: list[Evidence],
|
| 235 |
+
iteration: int = 0,
|
| 236 |
+
max_iterations: int = 10,
|
| 237 |
) -> JudgeAssessment:
|
| 238 |
"""
|
| 239 |
Assess evidence using HuggingFace Inference API.
|
|
|
|
| 262 |
|
| 263 |
# Format the user prompt
|
| 264 |
if evidence:
|
| 265 |
+
selected_evidence = await select_evidence_for_judge(evidence, question)
|
| 266 |
+
user_prompt = format_user_prompt(
|
| 267 |
+
question,
|
| 268 |
+
selected_evidence,
|
| 269 |
+
iteration,
|
| 270 |
+
max_iterations,
|
| 271 |
+
total_evidence_count=len(evidence),
|
| 272 |
+
)
|
| 273 |
else:
|
| 274 |
user_prompt = format_empty_evidence_prompt(question)
|
| 275 |
|
|
|
|
| 558 |
self,
|
| 559 |
question: str,
|
| 560 |
evidence: list[Evidence],
|
| 561 |
+
iteration: int = 0,
|
| 562 |
+
max_iterations: int = 10,
|
| 563 |
) -> JudgeAssessment:
|
| 564 |
"""Return assessment based on actual evidence (demo mode)."""
|
| 565 |
self.call_count += 1
|
|
@@ -40,12 +40,20 @@ class JudgeHandlerProtocol(Protocol):
|
|
| 40 |
and MockJudgeHandler.
|
| 41 |
"""
|
| 42 |
|
| 43 |
-
async def assess(
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
"""Assess whether collected evidence is sufficient.
|
| 45 |
|
| 46 |
Args:
|
| 47 |
question: The original research question
|
| 48 |
evidence: List of evidence items to assess
|
|
|
|
|
|
|
| 49 |
|
| 50 |
Returns:
|
| 51 |
JudgeAssessment with sufficiency determination and next steps
|
|
|
|
| 40 |
and MockJudgeHandler.
|
| 41 |
"""
|
| 42 |
|
| 43 |
+
async def assess(
|
| 44 |
+
self,
|
| 45 |
+
question: str,
|
| 46 |
+
evidence: list[Evidence],
|
| 47 |
+
iteration: int = 0,
|
| 48 |
+
max_iterations: int = 10,
|
| 49 |
+
) -> JudgeAssessment:
|
| 50 |
"""Assess whether collected evidence is sufficient.
|
| 51 |
|
| 52 |
Args:
|
| 53 |
question: The original research question
|
| 54 |
evidence: List of evidence items to assess
|
| 55 |
+
iteration: Current iteration number
|
| 56 |
+
max_iterations: Maximum allowed iterations
|
| 57 |
|
| 58 |
Returns:
|
| 59 |
JudgeAssessment with sufficiency determination and next steps
|
|
@@ -12,7 +12,7 @@ from __future__ import annotations
|
|
| 12 |
|
| 13 |
import asyncio
|
| 14 |
from collections.abc import AsyncGenerator
|
| 15 |
-
from typing import TYPE_CHECKING, Any
|
| 16 |
|
| 17 |
import structlog
|
| 18 |
|
|
@@ -42,6 +42,18 @@ class Orchestrator:
|
|
| 42 |
Microsoft Agent Framework.
|
| 43 |
"""
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
def __init__(
|
| 46 |
self,
|
| 47 |
search_handler: SearchHandlerProtocol,
|
|
@@ -100,6 +112,7 @@ class Orchestrator:
|
|
| 100 |
try:
|
| 101 |
# Deduplicate using semantic similarity
|
| 102 |
unique_evidence: list[Evidence] = await embeddings.deduplicate(evidence, threshold=0.85)
|
|
|
|
| 103 |
logger.info(
|
| 104 |
"Deduplicated evidence",
|
| 105 |
before=len(evidence),
|
|
@@ -153,6 +166,65 @@ class Orchestrator:
|
|
| 153 |
iteration=iteration,
|
| 154 |
)
|
| 155 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 156 |
async def run(self, query: str) -> AsyncGenerator[AgentEvent, None]: # noqa: PLR0915
|
| 157 |
"""
|
| 158 |
Run the agent loop for a query.
|
|
@@ -252,7 +324,9 @@ class Orchestrator:
|
|
| 252 |
)
|
| 253 |
|
| 254 |
try:
|
| 255 |
-
assessment = await self.judge.assess(
|
|
|
|
|
|
|
| 256 |
|
| 257 |
yield AgentEvent(
|
| 258 |
type="judge_complete",
|
|
@@ -279,15 +353,37 @@ class Orchestrator:
|
|
| 279 |
}
|
| 280 |
)
|
| 281 |
|
| 282 |
-
# === DECISION PHASE ===
|
| 283 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 284 |
# Optional Analysis Phase
|
| 285 |
async for event in self._run_analysis_phase(query, all_evidence, iteration):
|
| 286 |
yield event
|
| 287 |
|
| 288 |
yield AgentEvent(
|
| 289 |
type="synthesizing",
|
| 290 |
-
message="Evidence sufficient! Preparing synthesis...",
|
| 291 |
iteration=iteration,
|
| 292 |
)
|
| 293 |
|
|
@@ -300,6 +396,7 @@ class Orchestrator:
|
|
| 300 |
data={
|
| 301 |
"evidence_count": len(all_evidence),
|
| 302 |
"iterations": iteration,
|
|
|
|
| 303 |
"drug_candidates": assessment.details.drug_candidates,
|
| 304 |
"key_findings": assessment.details.key_findings,
|
| 305 |
},
|
|
@@ -317,10 +414,11 @@ class Orchestrator:
|
|
| 317 |
yield AgentEvent(
|
| 318 |
type="looping",
|
| 319 |
message=(
|
| 320 |
-
f"
|
| 321 |
-
f"
|
|
|
|
| 322 |
),
|
| 323 |
-
data={"next_queries": current_queries},
|
| 324 |
iteration=iteration,
|
| 325 |
)
|
| 326 |
|
|
@@ -410,36 +508,93 @@ class Orchestrator:
|
|
| 410 |
evidence: list[Evidence],
|
| 411 |
) -> str:
|
| 412 |
"""
|
| 413 |
-
Generate a
|
| 414 |
|
| 415 |
-
|
| 416 |
-
|
| 417 |
-
|
|
|
|
|
|
|
| 418 |
|
| 419 |
-
|
| 420 |
-
Formatted partial synthesis as markdown
|
| 421 |
"""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 422 |
citations = "\n".join(
|
| 423 |
[
|
| 424 |
-
f"{i + 1}. [{e.citation.title}]({e.citation.url})
|
|
|
|
| 425 |
for i, e in enumerate(evidence[:10])
|
| 426 |
]
|
| 427 |
)
|
| 428 |
|
| 429 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 430 |
|
| 431 |
-
|
|
|
|
|
|
|
| 432 |
{query}
|
| 433 |
|
| 434 |
### Status
|
| 435 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 436 |
|
| 437 |
-
###
|
| 438 |
-
|
| 439 |
|
| 440 |
-
### Citations
|
| 441 |
{citations}
|
| 442 |
|
| 443 |
---
|
| 444 |
-
*
|
|
|
|
|
|
|
|
|
|
| 445 |
"""
|
|
|
|
| 12 |
|
| 13 |
import asyncio
|
| 14 |
from collections.abc import AsyncGenerator
|
| 15 |
+
from typing import TYPE_CHECKING, Any, ClassVar
|
| 16 |
|
| 17 |
import structlog
|
| 18 |
|
|
|
|
| 42 |
Microsoft Agent Framework.
|
| 43 |
"""
|
| 44 |
|
| 45 |
+
# Termination thresholds (code-enforced, not LLM-decided)
|
| 46 |
+
TERMINATION_CRITERIA: ClassVar[dict[str, float]] = {
|
| 47 |
+
"min_combined_score": 12.0, # mechanism + clinical >= 12
|
| 48 |
+
"min_score_with_volume": 10.0, # >= 10 if 50+ sources
|
| 49 |
+
"min_evidence_for_volume": 50.0, # Priority 3: evidence count threshold
|
| 50 |
+
"late_iteration_threshold": 8.0, # >= 8 in iterations 8+
|
| 51 |
+
"max_evidence_threshold": 100.0, # Force synthesis with 100+ sources
|
| 52 |
+
"emergency_iteration": 8.0, # Last 2 iterations = emergency mode
|
| 53 |
+
"min_confidence": 0.5, # Minimum confidence for emergency synthesis
|
| 54 |
+
"min_evidence_for_emergency": 30.0, # Priority 6: min evidence for emergency
|
| 55 |
+
}
|
| 56 |
+
|
| 57 |
def __init__(
|
| 58 |
self,
|
| 59 |
search_handler: SearchHandlerProtocol,
|
|
|
|
| 112 |
try:
|
| 113 |
# Deduplicate using semantic similarity
|
| 114 |
unique_evidence: list[Evidence] = await embeddings.deduplicate(evidence, threshold=0.85)
|
| 115 |
+
|
| 116 |
logger.info(
|
| 117 |
"Deduplicated evidence",
|
| 118 |
before=len(evidence),
|
|
|
|
| 166 |
iteration=iteration,
|
| 167 |
)
|
| 168 |
|
| 169 |
+
def _should_synthesize(
|
| 170 |
+
self,
|
| 171 |
+
assessment: JudgeAssessment,
|
| 172 |
+
iteration: int,
|
| 173 |
+
max_iterations: int,
|
| 174 |
+
evidence_count: int,
|
| 175 |
+
) -> tuple[bool, str]:
|
| 176 |
+
"""
|
| 177 |
+
Code-enforced synthesis decision.
|
| 178 |
+
|
| 179 |
+
Returns (should_synthesize, reason).
|
| 180 |
+
"""
|
| 181 |
+
combined_score = (
|
| 182 |
+
assessment.details.mechanism_score + assessment.details.clinical_evidence_score
|
| 183 |
+
)
|
| 184 |
+
has_drug_candidates = len(assessment.details.drug_candidates) > 0
|
| 185 |
+
confidence = assessment.confidence
|
| 186 |
+
|
| 187 |
+
# Priority 1: LLM explicitly says sufficient with good scores
|
| 188 |
+
if assessment.sufficient and assessment.recommendation == "synthesize":
|
| 189 |
+
if combined_score >= 10:
|
| 190 |
+
return True, "judge_approved"
|
| 191 |
+
|
| 192 |
+
# Priority 2: High scores with drug candidates
|
| 193 |
+
if (
|
| 194 |
+
combined_score >= self.TERMINATION_CRITERIA["min_combined_score"]
|
| 195 |
+
and has_drug_candidates
|
| 196 |
+
):
|
| 197 |
+
return True, "high_scores_with_candidates"
|
| 198 |
+
|
| 199 |
+
# Priority 3: Good scores with high evidence volume
|
| 200 |
+
if (
|
| 201 |
+
combined_score >= self.TERMINATION_CRITERIA["min_score_with_volume"]
|
| 202 |
+
and evidence_count >= self.TERMINATION_CRITERIA["min_evidence_for_volume"]
|
| 203 |
+
):
|
| 204 |
+
return True, "good_scores_high_volume"
|
| 205 |
+
|
| 206 |
+
# Priority 4: Late iteration with acceptable scores (diminishing returns)
|
| 207 |
+
is_late_iteration = iteration >= max_iterations - 2
|
| 208 |
+
if (
|
| 209 |
+
is_late_iteration
|
| 210 |
+
and combined_score >= self.TERMINATION_CRITERIA["late_iteration_threshold"]
|
| 211 |
+
):
|
| 212 |
+
return True, "late_iteration_acceptable"
|
| 213 |
+
|
| 214 |
+
# Priority 5: Very high evidence count (enough to synthesize something)
|
| 215 |
+
if evidence_count >= self.TERMINATION_CRITERIA["max_evidence_threshold"]:
|
| 216 |
+
return True, "max_evidence_reached"
|
| 217 |
+
|
| 218 |
+
# Priority 6: Emergency synthesis (avoid garbage output)
|
| 219 |
+
if (
|
| 220 |
+
is_late_iteration
|
| 221 |
+
and evidence_count >= self.TERMINATION_CRITERIA["min_evidence_for_emergency"]
|
| 222 |
+
and confidence >= self.TERMINATION_CRITERIA["min_confidence"]
|
| 223 |
+
):
|
| 224 |
+
return True, "emergency_synthesis"
|
| 225 |
+
|
| 226 |
+
return False, "continue_searching"
|
| 227 |
+
|
| 228 |
async def run(self, query: str) -> AsyncGenerator[AgentEvent, None]: # noqa: PLR0915
|
| 229 |
"""
|
| 230 |
Run the agent loop for a query.
|
|
|
|
| 324 |
)
|
| 325 |
|
| 326 |
try:
|
| 327 |
+
assessment = await self.judge.assess(
|
| 328 |
+
query, all_evidence, iteration, self.config.max_iterations
|
| 329 |
+
)
|
| 330 |
|
| 331 |
yield AgentEvent(
|
| 332 |
type="judge_complete",
|
|
|
|
| 353 |
}
|
| 354 |
)
|
| 355 |
|
| 356 |
+
# === DECISION PHASE (Code-Enforced) ===
|
| 357 |
+
should_synth, reason = self._should_synthesize(
|
| 358 |
+
assessment=assessment,
|
| 359 |
+
iteration=iteration,
|
| 360 |
+
max_iterations=self.config.max_iterations,
|
| 361 |
+
evidence_count=len(all_evidence),
|
| 362 |
+
)
|
| 363 |
+
|
| 364 |
+
logger.info(
|
| 365 |
+
"Synthesis decision",
|
| 366 |
+
should_synthesize=should_synth,
|
| 367 |
+
reason=reason,
|
| 368 |
+
iteration=iteration,
|
| 369 |
+
combined_score=assessment.details.mechanism_score
|
| 370 |
+
+ assessment.details.clinical_evidence_score,
|
| 371 |
+
evidence_count=len(all_evidence),
|
| 372 |
+
confidence=assessment.confidence,
|
| 373 |
+
)
|
| 374 |
+
|
| 375 |
+
if should_synth:
|
| 376 |
+
# Log synthesis trigger reason for debugging
|
| 377 |
+
if reason != "judge_approved":
|
| 378 |
+
logger.info(f"Code-enforced synthesis triggered: {reason}")
|
| 379 |
+
|
| 380 |
# Optional Analysis Phase
|
| 381 |
async for event in self._run_analysis_phase(query, all_evidence, iteration):
|
| 382 |
yield event
|
| 383 |
|
| 384 |
yield AgentEvent(
|
| 385 |
type="synthesizing",
|
| 386 |
+
message=f"Evidence sufficient ({reason})! Preparing synthesis...",
|
| 387 |
iteration=iteration,
|
| 388 |
)
|
| 389 |
|
|
|
|
| 396 |
data={
|
| 397 |
"evidence_count": len(all_evidence),
|
| 398 |
"iterations": iteration,
|
| 399 |
+
"synthesis_reason": reason,
|
| 400 |
"drug_candidates": assessment.details.drug_candidates,
|
| 401 |
"key_findings": assessment.details.key_findings,
|
| 402 |
},
|
|
|
|
| 414 |
yield AgentEvent(
|
| 415 |
type="looping",
|
| 416 |
message=(
|
| 417 |
+
f"Gathering more evidence (scores: {assessment.details.mechanism_score}"
|
| 418 |
+
f"+{assessment.details.clinical_evidence_score}). "
|
| 419 |
+
f"Next: {', '.join(current_queries[:2])}..."
|
| 420 |
),
|
| 421 |
+
data={"next_queries": current_queries, "reason": reason},
|
| 422 |
iteration=iteration,
|
| 423 |
)
|
| 424 |
|
|
|
|
| 508 |
evidence: list[Evidence],
|
| 509 |
) -> str:
|
| 510 |
"""
|
| 511 |
+
Generate a REAL synthesis when max iterations reached.
|
| 512 |
|
| 513 |
+
Even when forced to stop, we should provide:
|
| 514 |
+
- Drug candidates (if any were found)
|
| 515 |
+
- Key findings
|
| 516 |
+
- Assessment scores
|
| 517 |
+
- Actionable citations
|
| 518 |
|
| 519 |
+
This is still better than a citation dump.
|
|
|
|
| 520 |
"""
|
| 521 |
+
# Extract data from last assessment if available
|
| 522 |
+
last_assessment = self.history[-1]["assessment"] if self.history else {}
|
| 523 |
+
details = last_assessment.get("details", {})
|
| 524 |
+
|
| 525 |
+
drug_candidates = details.get("drug_candidates", [])
|
| 526 |
+
key_findings = details.get("key_findings", [])
|
| 527 |
+
mechanism_score = details.get("mechanism_score", 0)
|
| 528 |
+
clinical_score = details.get("clinical_evidence_score", 0)
|
| 529 |
+
reasoning = last_assessment.get("reasoning", "Analysis incomplete due to iteration limit.")
|
| 530 |
+
|
| 531 |
+
# Format drug candidates
|
| 532 |
+
if drug_candidates:
|
| 533 |
+
drug_list = "\n".join([f"- **{d}**" for d in drug_candidates[:5]])
|
| 534 |
+
else:
|
| 535 |
+
drug_list = (
|
| 536 |
+
"- *No specific drug candidates identified in evidence*\n"
|
| 537 |
+
"- *Try a more specific query or add an API key for better analysis*"
|
| 538 |
+
)
|
| 539 |
+
|
| 540 |
+
# Format key findings
|
| 541 |
+
if key_findings:
|
| 542 |
+
findings_list = "\n".join([f"- {f}" for f in key_findings[:5]])
|
| 543 |
+
else:
|
| 544 |
+
findings_list = (
|
| 545 |
+
"- *Key findings require further analysis*\n"
|
| 546 |
+
"- *See citations below for relevant sources*"
|
| 547 |
+
)
|
| 548 |
+
|
| 549 |
+
# Format citations (top 10)
|
| 550 |
citations = "\n".join(
|
| 551 |
[
|
| 552 |
+
f"{i + 1}. [{e.citation.title}]({e.citation.url}) "
|
| 553 |
+
f"({e.citation.source.upper()}, {e.citation.date})"
|
| 554 |
for i, e in enumerate(evidence[:10])
|
| 555 |
]
|
| 556 |
)
|
| 557 |
|
| 558 |
+
combined_score = mechanism_score + clinical_score
|
| 559 |
+
mech_strength = (
|
| 560 |
+
"Strong" if mechanism_score >= 7 else "Moderate" if mechanism_score >= 4 else "Limited"
|
| 561 |
+
)
|
| 562 |
+
clin_strength = (
|
| 563 |
+
"Strong" if clinical_score >= 7 else "Moderate" if clinical_score >= 4 else "Limited"
|
| 564 |
+
)
|
| 565 |
+
comb_strength = "Sufficient" if combined_score >= 12 else "Partial"
|
| 566 |
|
| 567 |
+
return f"""## Drug Repurposing Analysis
|
| 568 |
+
|
| 569 |
+
### Research Question
|
| 570 |
{query}
|
| 571 |
|
| 572 |
### Status
|
| 573 |
+
Analysis based on {len(evidence)} sources across {len(self.history)} iterations.
|
| 574 |
+
Maximum iterations reached - results may be incomplete.
|
| 575 |
+
|
| 576 |
+
### Drug Candidates Identified
|
| 577 |
+
{drug_list}
|
| 578 |
+
|
| 579 |
+
### Key Findings
|
| 580 |
+
{findings_list}
|
| 581 |
+
|
| 582 |
+
### Evidence Quality Scores
|
| 583 |
+
| Criterion | Score | Interpretation |
|
| 584 |
+
|-----------|-------|----------------|
|
| 585 |
+
| Mechanism | {mechanism_score}/10 | {mech_strength} mechanistic evidence |
|
| 586 |
+
| Clinical | {clinical_score}/10 | {clin_strength} clinical support |
|
| 587 |
+
| Combined | {combined_score}/20 | {comb_strength} for synthesis |
|
| 588 |
|
| 589 |
+
### Analysis Summary
|
| 590 |
+
{reasoning}
|
| 591 |
|
| 592 |
+
### Top Citations ({len(evidence)} sources total)
|
| 593 |
{citations}
|
| 594 |
|
| 595 |
---
|
| 596 |
+
*For more complete analysis:*
|
| 597 |
+
- *Add an OpenAI or Anthropic API key for enhanced LLM analysis*
|
| 598 |
+
- *Try a more specific query (e.g., include drug names)*
|
| 599 |
+
- *Use Advanced mode for multi-agent research*
|
| 600 |
"""
|
|
@@ -4,10 +4,16 @@ from src.utils.models import Evidence
|
|
| 4 |
|
| 5 |
SYSTEM_PROMPT = """You are an expert drug repurposing research judge.
|
| 6 |
|
| 7 |
-
Your task is to
|
| 8 |
-
|
|
|
|
| 9 |
|
| 10 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
|
| 12 |
1. **Mechanism Score (0-10)**: How well does the evidence explain the biological mechanism?
|
| 13 |
- 0-3: No clear mechanism, speculative
|
|
@@ -19,59 +25,123 @@ recommend drug candidates for a given condition.
|
|
| 19 |
- 4-6: Preclinical or early clinical data
|
| 20 |
- 7-10: Strong clinical evidence (trials, meta-analyses)
|
| 21 |
|
| 22 |
-
3. **
|
| 23 |
-
-
|
| 24 |
-
-
|
| 25 |
-
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
"""
|
| 34 |
|
|
|
|
|
|
|
| 35 |
|
| 36 |
-
def
|
|
|
|
|
|
|
|
|
|
|
|
|
| 37 |
"""
|
| 38 |
-
|
| 39 |
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
-
|
| 45 |
-
|
| 46 |
"""
|
|
|
|
| 47 |
max_content_len = 1500
|
| 48 |
|
| 49 |
def format_single_evidence(i: int, e: Evidence) -> str:
|
| 50 |
content = e.content
|
| 51 |
if len(content) > max_content_len:
|
| 52 |
content = content[:max_content_len] + "..."
|
| 53 |
-
|
| 54 |
return (
|
| 55 |
f"### Evidence {i + 1}\n"
|
| 56 |
f"**Source**: {e.citation.source.upper()} - {e.citation.title}\n"
|
| 57 |
f"**URL**: {e.citation.url}\n"
|
| 58 |
-
f"**Date**: {e.citation.date}\n"
|
| 59 |
f"**Content**:\n{content}"
|
| 60 |
)
|
| 61 |
|
| 62 |
evidence_text = "\n\n".join([format_single_evidence(i, e) for i, e in enumerate(evidence)])
|
| 63 |
|
| 64 |
-
|
|
|
|
| 65 |
{question}
|
| 66 |
|
| 67 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 68 |
|
| 69 |
{evidence_text}
|
| 70 |
|
| 71 |
## Your Task
|
| 72 |
|
| 73 |
-
|
| 74 |
-
|
|
|
|
|
|
|
|
|
|
| 75 |
"""
|
| 76 |
|
| 77 |
|
|
|
|
| 4 |
|
| 5 |
SYSTEM_PROMPT = """You are an expert drug repurposing research judge.
|
| 6 |
|
| 7 |
+
Your task is to SCORE evidence from biomedical literature. You do NOT decide whether to
|
| 8 |
+
continue searching or synthesize - that decision is made by the orchestration system
|
| 9 |
+
based on your scores.
|
| 10 |
|
| 11 |
+
## Your Role: Scoring Only
|
| 12 |
+
|
| 13 |
+
You provide objective scores. The system decides next steps based on explicit thresholds.
|
| 14 |
+
This separation prevents bias in the decision-making process.
|
| 15 |
+
|
| 16 |
+
## Scoring Criteria
|
| 17 |
|
| 18 |
1. **Mechanism Score (0-10)**: How well does the evidence explain the biological mechanism?
|
| 19 |
- 0-3: No clear mechanism, speculative
|
|
|
|
| 25 |
- 4-6: Preclinical or early clinical data
|
| 26 |
- 7-10: Strong clinical evidence (trials, meta-analyses)
|
| 27 |
|
| 28 |
+
3. **Drug Candidates**: List SPECIFIC drug names mentioned in the evidence
|
| 29 |
+
- Only include drugs explicitly mentioned
|
| 30 |
+
- Do NOT hallucinate or infer drug names
|
| 31 |
+
- Include drug class if specific names aren't available (e.g., "SSRI antidepressants")
|
| 32 |
+
|
| 33 |
+
4. **Key Findings**: Extract 3-5 key findings from the evidence
|
| 34 |
+
- Focus on findings relevant to the research question
|
| 35 |
+
- Include mechanism insights and clinical outcomes
|
| 36 |
+
|
| 37 |
+
5. **Confidence (0.0-1.0)**: Your confidence in the scores
|
| 38 |
+
- Based on evidence quality and relevance
|
| 39 |
+
- Lower if evidence is tangential or low-quality
|
| 40 |
+
|
| 41 |
+
## Output Format
|
| 42 |
+
|
| 43 |
+
Return valid JSON with these fields:
|
| 44 |
+
- details.mechanism_score (int 0-10)
|
| 45 |
+
- details.mechanism_reasoning (string)
|
| 46 |
+
- details.clinical_evidence_score (int 0-10)
|
| 47 |
+
- details.clinical_reasoning (string)
|
| 48 |
+
- details.drug_candidates (list of strings)
|
| 49 |
+
- details.key_findings (list of strings)
|
| 50 |
+
- sufficient (boolean) - TRUE if scores suggest enough evidence
|
| 51 |
+
- confidence (float 0-1)
|
| 52 |
+
- recommendation ("continue" or "synthesize") - Your suggestion (system may override)
|
| 53 |
+
- next_search_queries (list) - If continuing, suggest FOCUSED queries
|
| 54 |
+
- reasoning (string)
|
| 55 |
+
|
| 56 |
+
## CRITICAL: Search Query Rules
|
| 57 |
+
|
| 58 |
+
When suggesting next_search_queries:
|
| 59 |
+
- STAY FOCUSED on the original research question
|
| 60 |
+
- Do NOT drift to tangential topics
|
| 61 |
+
- If question is about "female libido", do NOT suggest "bone health" or "muscle mass"
|
| 62 |
+
- Refine existing terms, don't explore random medical associations
|
| 63 |
"""
|
| 64 |
|
| 65 |
+
MAX_EVIDENCE_FOR_JUDGE = 30 # Keep under token limits
|
| 66 |
+
|
| 67 |
|
| 68 |
+
async def select_evidence_for_judge(
|
| 69 |
+
evidence: list[Evidence],
|
| 70 |
+
query: str,
|
| 71 |
+
max_items: int = MAX_EVIDENCE_FOR_JUDGE,
|
| 72 |
+
) -> list[Evidence]:
|
| 73 |
"""
|
| 74 |
+
Select diverse, relevant evidence for judge evaluation.
|
| 75 |
|
| 76 |
+
Implements RAG best practices:
|
| 77 |
+
- Diversity selection over recency-only
|
| 78 |
+
- Lost-in-the-middle mitigation
|
| 79 |
+
- Relevance re-ranking
|
| 80 |
+
"""
|
| 81 |
+
if len(evidence) <= max_items:
|
| 82 |
+
return evidence
|
| 83 |
+
|
| 84 |
+
try:
|
| 85 |
+
from src.utils.text_utils import select_diverse_evidence
|
| 86 |
+
|
| 87 |
+
# Use embedding-based diversity selection
|
| 88 |
+
return await select_diverse_evidence(evidence, n=max_items, query=query)
|
| 89 |
+
except ImportError:
|
| 90 |
+
# Fallback: mix of recent + early (lost-in-the-middle mitigation)
|
| 91 |
+
early = evidence[: max_items // 3] # First third
|
| 92 |
+
recent = evidence[-(max_items * 2 // 3) :] # Last two-thirds
|
| 93 |
+
return early + recent
|
| 94 |
+
|
| 95 |
+
|
| 96 |
+
def format_user_prompt(
|
| 97 |
+
question: str,
|
| 98 |
+
evidence: list[Evidence],
|
| 99 |
+
iteration: int = 0,
|
| 100 |
+
max_iterations: int = 10,
|
| 101 |
+
total_evidence_count: int | None = None,
|
| 102 |
+
) -> str:
|
| 103 |
+
"""
|
| 104 |
+
Format user prompt with selected evidence and iteration context.
|
| 105 |
|
| 106 |
+
NOTE: Evidence should be pre-selected using select_evidence_for_judge().
|
| 107 |
+
This function assumes evidence is already capped.
|
| 108 |
"""
|
| 109 |
+
total_count = total_evidence_count or len(evidence)
|
| 110 |
max_content_len = 1500
|
| 111 |
|
| 112 |
def format_single_evidence(i: int, e: Evidence) -> str:
|
| 113 |
content = e.content
|
| 114 |
if len(content) > max_content_len:
|
| 115 |
content = content[:max_content_len] + "..."
|
|
|
|
| 116 |
return (
|
| 117 |
f"### Evidence {i + 1}\n"
|
| 118 |
f"**Source**: {e.citation.source.upper()} - {e.citation.title}\n"
|
| 119 |
f"**URL**: {e.citation.url}\n"
|
|
|
|
| 120 |
f"**Content**:\n{content}"
|
| 121 |
)
|
| 122 |
|
| 123 |
evidence_text = "\n\n".join([format_single_evidence(i, e) for i, e in enumerate(evidence)])
|
| 124 |
|
| 125 |
+
# Lost-in-the-middle mitigation: put critical context at START and END
|
| 126 |
+
return f"""## Research Question (IMPORTANT - stay focused on this)
|
| 127 |
{question}
|
| 128 |
|
| 129 |
+
## Search Progress
|
| 130 |
+
- **Iteration**: {iteration}/{max_iterations}
|
| 131 |
+
- **Total evidence collected**: {total_count} sources
|
| 132 |
+
- **Evidence shown below**: {len(evidence)} diverse sources (selected for relevance)
|
| 133 |
+
|
| 134 |
+
## Available Evidence
|
| 135 |
|
| 136 |
{evidence_text}
|
| 137 |
|
| 138 |
## Your Task
|
| 139 |
|
| 140 |
+
Score this evidence for drug repurposing potential. Provide ONLY scores and extracted data.
|
| 141 |
+
DO NOT decide "synthesize" vs "continue" - that decision is made by the system.
|
| 142 |
+
|
| 143 |
+
## REMINDER: Original Question (stay focused)
|
| 144 |
+
{question}
|
| 145 |
"""
|
| 146 |
|
| 147 |
|
|
@@ -39,7 +39,7 @@ def mock_judge_handler():
|
|
| 39 |
"""Return a mock judge that always says 'synthesize'."""
|
| 40 |
mock = MagicMock()
|
| 41 |
|
| 42 |
-
async def mock_assess(question, evidence):
|
| 43 |
return JudgeAssessment(
|
| 44 |
sufficient=True,
|
| 45 |
confidence=0.9,
|
|
|
|
| 39 |
"""Return a mock judge that always says 'synthesize'."""
|
| 40 |
mock = MagicMock()
|
| 41 |
|
| 42 |
+
async def mock_assess(question, evidence, iteration=1, max_iterations=10):
|
| 43 |
return JudgeAssessment(
|
| 44 |
sufficient=True,
|
| 45 |
confidence=0.9,
|
|
@@ -0,0 +1,147 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from unittest.mock import AsyncMock
|
| 2 |
+
|
| 3 |
+
import pytest
|
| 4 |
+
|
| 5 |
+
from src.orchestrators.simple import Orchestrator
|
| 6 |
+
from src.utils.models import (
|
| 7 |
+
AssessmentDetails,
|
| 8 |
+
Citation,
|
| 9 |
+
Evidence,
|
| 10 |
+
JudgeAssessment,
|
| 11 |
+
OrchestratorConfig,
|
| 12 |
+
SearchResult,
|
| 13 |
+
)
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
def make_evidence(title: str) -> Evidence:
|
| 17 |
+
return Evidence(
|
| 18 |
+
content="content",
|
| 19 |
+
citation=Citation(title=title, url="http://test.com", date="2025", source="pubmed"),
|
| 20 |
+
)
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
@pytest.mark.integration
|
| 24 |
+
@pytest.mark.asyncio
|
| 25 |
+
async def test_simple_mode_synthesizes_before_max_iterations():
|
| 26 |
+
"""Verify simple mode produces useful output with mocked judge."""
|
| 27 |
+
# Mock search to return evidence
|
| 28 |
+
mock_search = AsyncMock()
|
| 29 |
+
mock_search.execute.return_value = SearchResult(
|
| 30 |
+
query="test query",
|
| 31 |
+
evidence=[make_evidence(f"Paper {i}") for i in range(5)],
|
| 32 |
+
errors=[],
|
| 33 |
+
sources_searched=["pubmed"],
|
| 34 |
+
total_found=5,
|
| 35 |
+
)
|
| 36 |
+
|
| 37 |
+
# Mock judge to return GOOD scores eventually
|
| 38 |
+
# We can use MockJudgeHandler or a pure mock. Let's use a pure mock to control scores precisely.
|
| 39 |
+
mock_judge = AsyncMock()
|
| 40 |
+
|
| 41 |
+
# Iteration 1: Low scores
|
| 42 |
+
assess_1 = JudgeAssessment(
|
| 43 |
+
details=AssessmentDetails(
|
| 44 |
+
mechanism_score=2,
|
| 45 |
+
mechanism_reasoning="reasoning is sufficient for valid model",
|
| 46 |
+
clinical_evidence_score=2,
|
| 47 |
+
clinical_reasoning="reasoning is sufficient for valid model",
|
| 48 |
+
drug_candidates=[],
|
| 49 |
+
key_findings=[],
|
| 50 |
+
),
|
| 51 |
+
sufficient=False,
|
| 52 |
+
confidence=0.5,
|
| 53 |
+
recommendation="continue",
|
| 54 |
+
next_search_queries=["q2"],
|
| 55 |
+
reasoning="need more evidence to support conclusions about this topic",
|
| 56 |
+
)
|
| 57 |
+
|
| 58 |
+
# Iteration 2: High scores (should trigger synthesis)
|
| 59 |
+
assess_2 = JudgeAssessment(
|
| 60 |
+
details=AssessmentDetails(
|
| 61 |
+
mechanism_score=8,
|
| 62 |
+
mechanism_reasoning="reasoning is sufficient for valid model",
|
| 63 |
+
clinical_evidence_score=7,
|
| 64 |
+
clinical_reasoning="reasoning is sufficient for valid model",
|
| 65 |
+
drug_candidates=["MagicDrug"],
|
| 66 |
+
key_findings=["It works"],
|
| 67 |
+
),
|
| 68 |
+
sufficient=False, # Judge is conservative
|
| 69 |
+
confidence=0.9,
|
| 70 |
+
recommendation="continue", # Judge still says continue (simulating bias)
|
| 71 |
+
next_search_queries=[],
|
| 72 |
+
reasoning="good scores but maybe more evidence needed technically",
|
| 73 |
+
)
|
| 74 |
+
|
| 75 |
+
mock_judge.assess.side_effect = [assess_1, assess_2]
|
| 76 |
+
|
| 77 |
+
orchestrator = Orchestrator(
|
| 78 |
+
search_handler=mock_search,
|
| 79 |
+
judge_handler=mock_judge,
|
| 80 |
+
config=OrchestratorConfig(max_iterations=5),
|
| 81 |
+
)
|
| 82 |
+
|
| 83 |
+
events = []
|
| 84 |
+
async for event in orchestrator.run("test query"):
|
| 85 |
+
events.append(event)
|
| 86 |
+
if event.type == "complete":
|
| 87 |
+
break
|
| 88 |
+
|
| 89 |
+
# Must have synthesis with drug candidates
|
| 90 |
+
complete_events = [e for e in events if e.type == "complete"]
|
| 91 |
+
assert len(complete_events) == 1
|
| 92 |
+
complete_event = complete_events[0]
|
| 93 |
+
|
| 94 |
+
assert "MagicDrug" in complete_event.message
|
| 95 |
+
assert "Drug Candidates" in complete_event.message
|
| 96 |
+
assert complete_event.data.get("synthesis_reason") == "high_scores_with_candidates"
|
| 97 |
+
assert complete_event.iteration == 2 # Should stop at it 2
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
@pytest.mark.integration
|
| 101 |
+
@pytest.mark.asyncio
|
| 102 |
+
async def test_partial_synthesis_generation():
|
| 103 |
+
"""Verify partial synthesis includes drug candidates even if max iterations reached."""
|
| 104 |
+
mock_search = AsyncMock()
|
| 105 |
+
mock_search.execute.return_value = SearchResult(
|
| 106 |
+
query="test", evidence=[], errors=[], sources_searched=["pubmed"], total_found=0
|
| 107 |
+
)
|
| 108 |
+
|
| 109 |
+
mock_judge = AsyncMock()
|
| 110 |
+
# Always return low scores but WITH candidates
|
| 111 |
+
# Scores 3+3 = 6 < 8 (late threshold), so it should NOT synthesize early
|
| 112 |
+
mock_judge.assess.return_value = JudgeAssessment(
|
| 113 |
+
details=AssessmentDetails(
|
| 114 |
+
mechanism_score=3,
|
| 115 |
+
mechanism_reasoning="reasoning is sufficient for valid model",
|
| 116 |
+
clinical_evidence_score=3,
|
| 117 |
+
clinical_reasoning="reasoning is sufficient for valid model",
|
| 118 |
+
drug_candidates=["PartialDrug"],
|
| 119 |
+
key_findings=["Partial finding"],
|
| 120 |
+
),
|
| 121 |
+
sufficient=False,
|
| 122 |
+
confidence=0.5,
|
| 123 |
+
recommendation="continue",
|
| 124 |
+
next_search_queries=[],
|
| 125 |
+
reasoning="keep going to find more evidence about this topic please",
|
| 126 |
+
)
|
| 127 |
+
|
| 128 |
+
orchestrator = Orchestrator(
|
| 129 |
+
search_handler=mock_search,
|
| 130 |
+
judge_handler=mock_judge,
|
| 131 |
+
config=OrchestratorConfig(max_iterations=2),
|
| 132 |
+
)
|
| 133 |
+
|
| 134 |
+
events = []
|
| 135 |
+
async for event in orchestrator.run("test"):
|
| 136 |
+
events.append(event)
|
| 137 |
+
|
| 138 |
+
complete_events = [e for e in events if e.type == "complete"]
|
| 139 |
+
assert (
|
| 140 |
+
len(complete_events) == 1
|
| 141 |
+
), f"Expected exactly one complete event, got {len(complete_events)}"
|
| 142 |
+
complete_event = complete_events[0]
|
| 143 |
+
assert complete_event.data.get("max_reached") is True
|
| 144 |
+
|
| 145 |
+
# The output message should contain the drug candidate from the last assessment
|
| 146 |
+
assert "PartialDrug" in complete_event.message
|
| 147 |
+
assert "Maximum iterations reached" in complete_event.message
|
|
@@ -0,0 +1,104 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Literal
|
| 2 |
+
from unittest.mock import MagicMock
|
| 3 |
+
|
| 4 |
+
import pytest
|
| 5 |
+
|
| 6 |
+
from src.orchestrators.simple import Orchestrator
|
| 7 |
+
from src.utils.models import AssessmentDetails, JudgeAssessment
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
def make_assessment(
|
| 11 |
+
mechanism: int,
|
| 12 |
+
clinical: int,
|
| 13 |
+
drug_candidates: list[str],
|
| 14 |
+
sufficient: bool = False,
|
| 15 |
+
recommendation: Literal["continue", "synthesize"] = "continue",
|
| 16 |
+
confidence: float = 0.8,
|
| 17 |
+
) -> JudgeAssessment:
|
| 18 |
+
return JudgeAssessment(
|
| 19 |
+
details=AssessmentDetails(
|
| 20 |
+
mechanism_score=mechanism,
|
| 21 |
+
mechanism_reasoning="reasoning is sufficient for testing purposes",
|
| 22 |
+
clinical_evidence_score=clinical,
|
| 23 |
+
clinical_reasoning="reasoning is sufficient for testing purposes",
|
| 24 |
+
drug_candidates=drug_candidates,
|
| 25 |
+
key_findings=["finding"],
|
| 26 |
+
),
|
| 27 |
+
sufficient=sufficient,
|
| 28 |
+
confidence=confidence,
|
| 29 |
+
recommendation=recommendation,
|
| 30 |
+
next_search_queries=[],
|
| 31 |
+
reasoning="reasoning is sufficient for testing purposes",
|
| 32 |
+
)
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
@pytest.fixture
|
| 36 |
+
def orchestrator():
|
| 37 |
+
search = MagicMock()
|
| 38 |
+
judge = MagicMock()
|
| 39 |
+
return Orchestrator(search, judge)
|
| 40 |
+
|
| 41 |
+
|
| 42 |
+
@pytest.mark.unit
|
| 43 |
+
def test_should_synthesize_high_scores(orchestrator):
|
| 44 |
+
"""High scores with drug candidates triggers synthesis."""
|
| 45 |
+
assessment = make_assessment(mechanism=7, clinical=6, drug_candidates=["Metformin"])
|
| 46 |
+
|
| 47 |
+
# Access the private method via name mangling or just call it if it was public.
|
| 48 |
+
# Since I made it private _should_synthesize, I access it directly.
|
| 49 |
+
should_synth, reason = orchestrator._should_synthesize(
|
| 50 |
+
assessment, iteration=3, max_iterations=10, evidence_count=50
|
| 51 |
+
)
|
| 52 |
+
|
| 53 |
+
assert should_synth is True
|
| 54 |
+
assert reason == "high_scores_with_candidates"
|
| 55 |
+
|
| 56 |
+
|
| 57 |
+
@pytest.mark.unit
|
| 58 |
+
def test_should_synthesize_late_iteration(orchestrator):
|
| 59 |
+
"""Late iteration with acceptable scores triggers synthesis."""
|
| 60 |
+
assessment = make_assessment(mechanism=5, clinical=4, drug_candidates=[])
|
| 61 |
+
should_synth, reason = orchestrator._should_synthesize(
|
| 62 |
+
assessment, iteration=9, max_iterations=10, evidence_count=80
|
| 63 |
+
)
|
| 64 |
+
|
| 65 |
+
assert should_synth is True
|
| 66 |
+
assert reason in ["late_iteration_acceptable", "emergency_synthesis"]
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
@pytest.mark.unit
|
| 70 |
+
def test_should_not_synthesize_early_low_scores(orchestrator):
|
| 71 |
+
"""Early iteration with low scores continues searching."""
|
| 72 |
+
assessment = make_assessment(mechanism=3, clinical=2, drug_candidates=[])
|
| 73 |
+
should_synth, reason = orchestrator._should_synthesize(
|
| 74 |
+
assessment, iteration=2, max_iterations=10, evidence_count=20
|
| 75 |
+
)
|
| 76 |
+
|
| 77 |
+
assert should_synth is False
|
| 78 |
+
assert reason == "continue_searching"
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
@pytest.mark.unit
|
| 82 |
+
def test_judge_approved_overrides_all(orchestrator):
|
| 83 |
+
"""If judge explicitly says synthesize with good scores, do it."""
|
| 84 |
+
assessment = make_assessment(
|
| 85 |
+
mechanism=6, clinical=5, drug_candidates=[], sufficient=True, recommendation="synthesize"
|
| 86 |
+
)
|
| 87 |
+
should_synth, reason = orchestrator._should_synthesize(
|
| 88 |
+
assessment, iteration=2, max_iterations=10, evidence_count=20
|
| 89 |
+
)
|
| 90 |
+
|
| 91 |
+
assert should_synth is True
|
| 92 |
+
assert reason == "judge_approved"
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
@pytest.mark.unit
|
| 96 |
+
def test_max_evidence_threshold(orchestrator):
|
| 97 |
+
"""Force synthesis if we have tons of evidence."""
|
| 98 |
+
assessment = make_assessment(mechanism=2, clinical=2, drug_candidates=[])
|
| 99 |
+
should_synth, reason = orchestrator._should_synthesize(
|
| 100 |
+
assessment, iteration=5, max_iterations=10, evidence_count=150
|
| 101 |
+
)
|
| 102 |
+
|
| 103 |
+
assert should_synth is True
|
| 104 |
+
assert reason == "max_evidence_reached"
|
|
@@ -0,0 +1,61 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from unittest.mock import patch
|
| 2 |
+
|
| 3 |
+
import pytest
|
| 4 |
+
|
| 5 |
+
from src.prompts.judge import format_user_prompt, select_evidence_for_judge
|
| 6 |
+
from src.utils.models import Citation, Evidence
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
def make_evidence(title: str, content: str = "content") -> Evidence:
|
| 10 |
+
return Evidence(
|
| 11 |
+
content=content,
|
| 12 |
+
citation=Citation(title=title, url="http://test.com", date="2025", source="pubmed"),
|
| 13 |
+
)
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
@pytest.mark.unit
|
| 17 |
+
@pytest.mark.asyncio
|
| 18 |
+
async def test_evidence_selection_diverse():
|
| 19 |
+
"""Verify evidence selection includes early and recent items (fallback logic)."""
|
| 20 |
+
# Create enough evidence to trigger selection
|
| 21 |
+
evidence = [make_evidence(f"Paper {i}") for i in range(100)]
|
| 22 |
+
|
| 23 |
+
# Mock select_diverse_evidence to raise ImportError to trigger fallback logic
|
| 24 |
+
with patch("src.utils.text_utils.select_diverse_evidence", side_effect=ImportError):
|
| 25 |
+
selected = await select_evidence_for_judge(evidence, "test query", max_items=30)
|
| 26 |
+
|
| 27 |
+
assert len(selected) == 30
|
| 28 |
+
|
| 29 |
+
# Should include some early evidence (lost-in-the-middle mitigation)
|
| 30 |
+
titles = [e.citation.title for e in selected]
|
| 31 |
+
|
| 32 |
+
# Check for start (Paper 0..9) - using set intersection for clarity
|
| 33 |
+
early_papers = {f"Paper {i}" for i in range(10)}
|
| 34 |
+
has_early = any(title in early_papers for title in titles)
|
| 35 |
+
# Check for end (Paper 90..99)
|
| 36 |
+
late_papers = {f"Paper {i}" for i in range(90, 100)}
|
| 37 |
+
has_late = any(title in late_papers for title in titles)
|
| 38 |
+
|
| 39 |
+
assert has_early, "Should include early evidence"
|
| 40 |
+
assert has_late, "Should include recent evidence"
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
@pytest.mark.unit
|
| 44 |
+
def test_prompt_includes_question_at_edges():
|
| 45 |
+
"""Verify lost-in-the-middle mitigation in prompt formatting."""
|
| 46 |
+
evidence = [make_evidence("Test Paper")]
|
| 47 |
+
question = "CRITICAL RESEARCH QUESTION"
|
| 48 |
+
|
| 49 |
+
prompt = format_user_prompt(question, evidence, iteration=5, max_iterations=10)
|
| 50 |
+
|
| 51 |
+
# Question should appear at START and END of prompt
|
| 52 |
+
lines = prompt.split("\n")
|
| 53 |
+
|
| 54 |
+
# Check start (first few lines)
|
| 55 |
+
start_content = "\n".join(lines[:10])
|
| 56 |
+
assert question in start_content
|
| 57 |
+
|
| 58 |
+
# Check end (last few lines)
|
| 59 |
+
end_content = "\n".join(lines[-10:])
|
| 60 |
+
assert question in end_content
|
| 61 |
+
assert "REMINDER: Original Question" in end_content
|