File size: 16,001 Bytes
6d1bbc7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
"""Tests for CT LLM evaluation functions (src/negbiodb_ct/llm_eval.py)."""

import json

import pytest

from negbiodb_ct.llm_eval import (
    CT_EVIDENCE_KEYWORDS,
    CT_L2_REQUIRED_FIELDS,
    CT_L3_JUDGE_PROMPT,
    compute_all_ct_llm_metrics,
    evaluate_ct_l1,
    evaluate_ct_l2,
    evaluate_ct_l3,
    evaluate_ct_l4,
    parse_ct_l1_answer,
    parse_ct_l2_response,
    parse_ct_l3_judge_scores,
    parse_ct_l4_answer,
)


# ── CT-L1 Parser Tests ───────────────────────────────────────────────────


class TestParseCTL1Answer:
    def test_single_letter_a_through_e(self):
        for letter in "ABCDE":
            assert parse_ct_l1_answer(letter) == letter

    def test_case_insensitive(self):
        assert parse_ct_l1_answer("a") == "A"
        assert parse_ct_l1_answer("e") == "E"

    def test_answer_colon_format(self):
        assert parse_ct_l1_answer("Answer: E") == "E"
        assert parse_ct_l1_answer("Answer: B") == "B"

    def test_answer_is_format(self):
        assert parse_ct_l1_answer("The answer is C") == "C"

    def test_parenthesized(self):
        assert parse_ct_l1_answer("(D) Strategic discontinuation") == "D"

    def test_letter_dot_format(self):
        assert parse_ct_l1_answer("A. Safety issue") == "A"

    def test_letter_with_explanation(self):
        assert parse_ct_l1_answer("B\nThis trial failed to show efficacy") == "B"

    def test_empty_returns_none(self):
        assert parse_ct_l1_answer("") is None

    def test_no_valid_letter_returns_none(self):
        assert parse_ct_l1_answer("I don't know the answer") is None

    def test_e_category_recognized(self):
        """CT uses E (5-way) unlike DTI (4-way A-D)."""
        assert parse_ct_l1_answer("E") == "E"
        assert parse_ct_l1_answer("Answer: E") == "E"
        assert parse_ct_l1_answer("(E)") == "E"


# ── CT-L1 Evaluator Tests ───────────────────────────────────────────────


class TestEvaluateCTL1:
    def test_perfect_accuracy(self):
        preds = ["A", "B", "C", "D", "E"]
        golds = ["A", "B", "C", "D", "E"]
        result = evaluate_ct_l1(preds, golds)
        assert result["accuracy"] == 1.0
        assert result["parse_rate"] == 1.0

    def test_zero_accuracy(self):
        preds = ["B", "A", "D", "C", "A"]
        golds = ["A", "B", "C", "D", "E"]
        result = evaluate_ct_l1(preds, golds)
        assert result["accuracy"] == 0.0

    def test_per_class_accuracy(self):
        preds = ["A", "B", "C"]
        golds = ["A", "B", "D"]
        classes = ["safety", "efficacy", "strategic"]
        result = evaluate_ct_l1(preds, golds, gold_classes=classes)
        assert "per_class_accuracy" in result
        assert result["per_class_accuracy"]["safety"] == 1.0
        assert result["per_class_accuracy"]["strategic"] == 0.0

    def test_per_difficulty_accuracy(self):
        preds = ["A", "A", "B", "B"]
        golds = ["A", "B", "B", "A"]
        diffs = ["easy", "easy", "hard", "hard"]
        result = evaluate_ct_l1(preds, golds, difficulties=diffs)
        assert "per_difficulty_accuracy" in result
        assert result["per_difficulty_accuracy"]["easy"] == 0.5
        assert result["per_difficulty_accuracy"]["hard"] == 0.5

    def test_parse_failures(self):
        preds = ["A", "no valid response here at all", "C"]
        golds = ["A", "B", "C"]
        result = evaluate_ct_l1(preds, golds)
        assert result["n_valid"] == 2
        assert result["parse_rate"] == pytest.approx(2 / 3)
        assert result["accuracy"] == 1.0  # Both parsed correctly

    def test_empty_predictions(self):
        result = evaluate_ct_l1([], [])
        assert result["accuracy"] == 0.0
        assert result["n_total"] == 0

    def test_all_unparseable(self):
        preds = ["xyz", "hello", "???"]
        golds = ["A", "B", "C"]
        result = evaluate_ct_l1(preds, golds)
        assert result["accuracy"] == 0.0
        assert result["n_valid"] == 0


# ── CT-L2 Parser Tests ──────────────────────────────────────────────────


class TestParseCTL2Response:
    def test_valid_json(self):
        obj = {"failure_category": "efficacy", "failure_subcategory": "futility"}
        result = parse_ct_l2_response(json.dumps(obj))
        assert result == obj

    def test_json_with_code_fences(self):
        raw = '```json\n{"failure_category": "safety"}\n```'
        result = parse_ct_l2_response(raw)
        assert result["failure_category"] == "safety"

    def test_json_embedded_in_text(self):
        raw = 'Here is the result: {"failure_category": "enrollment"} as expected.'
        result = parse_ct_l2_response(raw)
        assert result["failure_category"] == "enrollment"

    def test_invalid_json(self):
        assert parse_ct_l2_response("not json at all") is None

    def test_partial_fields(self):
        obj = {"failure_category": "safety"}  # Missing other fields
        result = parse_ct_l2_response(json.dumps(obj))
        assert result is not None
        assert result["failure_category"] == "safety"


# ── CT-L2 Evaluator Tests ───────────────────────────────────────────────


class TestEvaluateCTL2:
    def test_perfect_schema_compliance(self):
        pred_obj = {f: "test" for f in CT_L2_REQUIRED_FIELDS}
        pred_obj["quantitative_evidence"] = True
        preds = [json.dumps(pred_obj)]
        golds = [{"gold_answer": "efficacy", "failure_category": "efficacy"}]
        result = evaluate_ct_l2(preds, golds)
        assert result["schema_compliance"] == 1.0
        assert result["parse_rate"] == 1.0

    def test_category_accuracy(self):
        pred_obj = {"failure_category": "safety", "failure_subcategory": "toxicity",
                     "affected_system": "liver", "severity_indicator": "severe",
                     "quantitative_evidence": False, "decision_maker": "dsmb",
                     "patient_impact": "hepatic injury"}
        preds = [json.dumps(pred_obj)]
        golds = [{"gold_answer": "safety"}]
        result = evaluate_ct_l2(preds, golds)
        assert result["category_accuracy"] == 1.0

    def test_wrong_category(self):
        pred_obj = {"failure_category": "efficacy"}
        preds = [json.dumps(pred_obj)]
        golds = [{"gold_answer": "safety"}]
        result = evaluate_ct_l2(preds, golds)
        assert result["category_accuracy"] == 0.0

    def test_parse_rate(self):
        preds = ['{"failure_category": "safety"}', "not json", '{"failure_category": "efficacy"}']
        golds = [{"gold_answer": "safety"}, {"gold_answer": "efficacy"}, {"gold_answer": "efficacy"}]
        result = evaluate_ct_l2(preds, golds)
        assert result["parse_rate"] == pytest.approx(2 / 3)

    def test_empty_predictions(self):
        result = evaluate_ct_l2([], [])
        assert result["n_total"] == 0


# ── CT-L3 Judge Score Parser Tests ───────────────────────────────────────


class TestParseCTL3JudgeScores:
    def test_valid_scores(self):
        resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5, "specificity": 2})
        scores = parse_ct_l3_judge_scores(resp)
        assert scores == {"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0}

    def test_out_of_range(self):
        resp = json.dumps({"accuracy": 6, "reasoning": 0, "completeness": 3, "specificity": 3})
        scores = parse_ct_l3_judge_scores(resp)
        assert scores is None  # 6 and 0 are out of range

    def test_missing_dimension(self):
        resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5})
        scores = parse_ct_l3_judge_scores(resp)
        assert scores is None  # specificity missing

    def test_invalid_json(self):
        scores = parse_ct_l3_judge_scores("not json")
        assert scores is None


# ── CT-L3 Evaluator Tests ───────────────────────────────────────────────


class TestEvaluateCTL3:
    def test_aggregation(self):
        scores = [
            {"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0},
            {"accuracy": 2.0, "reasoning": 5.0, "completeness": 3.0, "specificity": 4.0},
        ]
        result = evaluate_ct_l3(scores)
        assert result["accuracy"]["mean"] == pytest.approx(3.0)
        assert result["reasoning"]["mean"] == pytest.approx(4.0)
        assert result["overall"]["mean"] == pytest.approx(3.5)
        assert result["n_valid"] == 2

    def test_none_handling(self):
        scores = [
            {"accuracy": 4.0, "reasoning": 3.0, "completeness": 5.0, "specificity": 2.0},
            None,
        ]
        result = evaluate_ct_l3(scores)
        assert result["n_valid"] == 1
        assert result["n_total"] == 2

    def test_all_none(self):
        result = evaluate_ct_l3([None, None])
        assert result["n_valid"] == 0
        assert result["accuracy"]["mean"] == 0.0

    def test_empty(self):
        result = evaluate_ct_l3([])
        assert result["n_valid"] == 0


# ── CT-L4 Parser Tests ──────────────────────────────────────────────────


class TestParseCTL4Answer:
    def test_tested(self):
        answer, evidence = parse_ct_l4_answer("tested\nNCT01234567 completed in 2020")
        assert answer == "tested"
        assert "NCT01234567" in evidence

    def test_untested(self):
        answer, evidence = parse_ct_l4_answer("untested\nNo registered trials found")
        assert answer == "untested"
        assert "No registered" in evidence

    def test_not_tested_variant(self):
        answer, _ = parse_ct_l4_answer("not tested\nReasoning...")
        assert answer == "untested"

    def test_not_been_tested_variant(self):
        answer, _ = parse_ct_l4_answer("This combination has not been tested\nEvidence...")
        assert answer == "untested"

    def test_never_been_tested_variant(self):
        answer, _ = parse_ct_l4_answer("This drug-disease combination has never been tested in a trial.")
        assert answer == "untested"

    def test_no_evidence(self):
        answer, evidence = parse_ct_l4_answer("tested")
        assert answer == "tested"
        assert evidence is None

    def test_empty(self):
        answer, evidence = parse_ct_l4_answer("")
        assert answer is None
        assert evidence is None


# ── CT-L4 Evaluator Tests ───────────────────────────────────────────────


class TestEvaluateCTL4:
    def test_perfect_accuracy(self):
        preds = ["tested\nNCT123", "untested\nNo trials"]
        golds = ["tested", "untested"]
        result = evaluate_ct_l4(preds, golds)
        assert result["accuracy"] == 1.0

    def test_temporal_pre_2020_post_2023(self):
        """CT uses pre_2020/post_2023, NOT DTI's pre_2023/post_2024."""
        preds = ["tested\nNCT001", "tested\nNCT002", "untested\nNone", "tested\nNCT003"]
        golds = ["tested", "tested", "untested", "untested"]
        temporal = ["pre_2020", "post_2023", "pre_2020", "post_2023"]
        result = evaluate_ct_l4(preds, golds, temporal_groups=temporal)
        # pre_2020: tested→tested (correct), untested→untested (correct) → 100%
        assert result["accuracy_pre_2020"] == 1.0
        # post_2023: tested→tested (correct), untested→tested (wrong) → 50%
        assert result["accuracy_post_2023"] == 0.5

    def test_contamination_flag(self):
        """Flag when pre_2020 accuracy exceeds post_2023 by >15%."""
        preds = ["tested\nA", "tested\nB", "tested\nC", "untested\nD",
                  "tested\nE", "tested\nF", "tested\nG", "tested\nH"]
        golds = ["tested", "tested", "tested", "untested",
                  "untested", "untested", "untested", "untested"]
        temporal = ["pre_2020", "pre_2020", "pre_2020", "pre_2020",
                     "post_2023", "post_2023", "post_2023", "post_2023"]
        result = evaluate_ct_l4(preds, golds, temporal)
        # pre_2020: 3 correct + 1 correct = 4/4 = 100%
        # post_2023: 0/4 = 0%
        assert result["contamination_flag"] is True
        assert result["contamination_gap"] == pytest.approx(1.0)

    def test_no_contamination(self):
        preds = ["tested\nA", "untested\nB"]
        golds = ["tested", "untested"]
        temporal = ["pre_2020", "post_2023"]
        result = evaluate_ct_l4(preds, golds, temporal)
        assert result["contamination_flag"] is False

    def test_evidence_citation_rate(self):
        """Evidence needs BOTH >50 chars AND domain keyword (AND logic)."""
        preds = [
            # >50 chars AND contains "nct" keyword → pass
            "tested\nTrial NCT01234567 demonstrated that the drug was effective in reducing primary endpoint with p-value 0.003",
            # >50 chars but NO keyword → fail
            "tested\nI think the drug was probably tested somewhere in a large randomized clinical study recently",
            # <50 chars but has keyword → fail
            "tested\nNCT01234567 showed results",
        ]
        golds = ["tested", "tested", "tested"]
        result = evaluate_ct_l4(preds, golds)
        # Only 1 of 3 passes both conditions
        assert result["evidence_citation_rate"] == pytest.approx(1 / 3)

    def test_ct_evidence_keywords(self):
        """CT-specific keywords differ from DTI."""
        assert "nct" in CT_EVIDENCE_KEYWORDS
        assert "clinicaltrials" in CT_EVIDENCE_KEYWORDS
        assert "fda" in CT_EVIDENCE_KEYWORDS
        assert "eudract" in CT_EVIDENCE_KEYWORDS
        # DTI keywords should NOT be here
        assert "chembl" not in CT_EVIDENCE_KEYWORDS
        assert "pubchem" not in CT_EVIDENCE_KEYWORDS

    def test_empty(self):
        result = evaluate_ct_l4([], [])
        assert result["accuracy"] == 0.0


# ── Dispatch Tests ───────────────────────────────────────────────────────


class TestDispatch:
    def test_ct_l1_dispatch(self):
        preds = ["A", "B"]
        gold = [{"gold_answer": "A", "gold_category": "safety", "difficulty": "easy"},
                {"gold_answer": "B", "gold_category": "efficacy", "difficulty": "hard"}]
        result = compute_all_ct_llm_metrics("ct-l1", preds, gold)
        assert result["accuracy"] == 1.0

    def test_ct_l4_dispatch(self):
        preds = ["tested\nNCT123", "untested\nNone"]
        gold = [{"gold_answer": "tested", "temporal_group": "pre_2020"},
                {"gold_answer": "untested", "temporal_group": "post_2023"}]
        result = compute_all_ct_llm_metrics("ct-l4", preds, gold)
        assert result["accuracy"] == 1.0

    def test_invalid_task_raises(self):
        with pytest.raises(ValueError, match="Unknown task"):
            compute_all_ct_llm_metrics("l1", ["A"], [{"gold_answer": "A"}])

    def test_ct_l2_dispatch(self):
        pred_obj = {"failure_category": "efficacy"}
        preds = [json.dumps(pred_obj)]
        gold = [{"gold_answer": "efficacy"}]
        result = compute_all_ct_llm_metrics("ct-l2", preds, gold)
        assert result["category_accuracy"] == 1.0

    def test_ct_l3_dispatch(self):
        resp = json.dumps({"accuracy": 4, "reasoning": 3, "completeness": 5, "specificity": 2})
        result = compute_all_ct_llm_metrics("ct-l3", [resp], [{}])
        assert result["n_valid"] == 1