File size: 3,342 Bytes
e5e4b98
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
# Copyright (c) Meta Platforms, Inc. and affiliates. All Rights Reserved

"""Script to run the evaluator offline given the GTs for SAC-Gold test set and SAM3 model prediction files.
It reports CGF1, IL_MCC, PM_F1 metrics for each subset of SAC-Gold test set.

Usage: python eval_sam3.py --gt-folder <folder_with_gts> --pred-folder <folder_with_predictions>
"""

import argparse
import os

from sam3.eval.cgf1_eval import CGF1Evaluator

# Relative file names for GT files for 7 SA-Co/Gold subsets

saco_gold_gts = {
    # MetaCLIP Captioner
    "metaclip_nps": [
        "gold_metaclip_merged_a_release_test.json",
        "gold_metaclip_merged_b_release_test.json",
        "gold_metaclip_merged_c_release_test.json",
    ],
    # SA-1B captioner
    "sa1b_nps": [
        "gold_sa1b_merged_a_release_test.json",
        "gold_sa1b_merged_b_release_test.json",
        "gold_sa1b_merged_c_release_test.json",
    ],
    # Crowded
    "crowded": [
        "gold_crowded_merged_a_release_test.json",
        "gold_crowded_merged_b_release_test.json",
        "gold_crowded_merged_c_release_test.json",
    ],
    # FG Food
    "fg_food": [
        "gold_fg_food_merged_a_release_test.json",
        "gold_fg_food_merged_b_release_test.json",
        "gold_fg_food_merged_c_release_test.json",
    ],
    # FG Sports
    "fg_sports_equipment": [
        "gold_fg_sports_equipment_merged_a_release_test.json",
        "gold_fg_sports_equipment_merged_b_release_test.json",
        "gold_fg_sports_equipment_merged_c_release_test.json",
    ],
    # Attributes
    "attributes": [
        "gold_attributes_merged_a_release_test.json",
        "gold_attributes_merged_b_release_test.json",
        "gold_attributes_merged_c_release_test.json",
    ],
    # Wiki common
    "wiki_common": [
        "gold_wiki_common_merged_a_release_test.json",
        "gold_wiki_common_merged_b_release_test.json",
        "gold_wiki_common_merged_c_release_test.json",
    ],
}


def main():
    parser = argparse.ArgumentParser()
    parser.add_argument(
        "-g",
        "--gt-folder",
        type=str,
        help="Path to the folder containing the ground truth json files.",
    )
    parser.add_argument(
        "-p",
        "--pred-folder",
        type=str,
        help="Path to the folder containing the predictions json files.",
    )
    args = parser.parse_args()

    results = ""

    for subset_name, gts in saco_gold_gts.items():
        print("Processing subset: ", subset_name)
        gt_paths = [os.path.join(args.gt_folder, gt) for gt in gts]
        evaluator = CGF1Evaluator(
            gt_path=gt_paths, verbose=True, iou_type="segm"
        )  # change to bbox if you want detection performance

        pred_path = os.path.join(
            args.pred_folder,
            f"gold_{subset_name}/dumps/gold_{subset_name}/coco_predictions_segm.json",
        )
        summary = evaluator.evaluate(pred_path)

        cgf1 = str(round(summary["cgF1_eval_segm_cgF1"] * 100, 2))
        il_mcc = str(round(summary["cgF1_eval_segm_IL_MCC"], 2))
        pmf1 = str(round(summary["cgF1_eval_segm_positive_micro_F1"] * 100, 2))
        final_str = f"{cgf1},{il_mcc},{pmf1}"
        results += subset_name + ": " + final_str + "\n"

    print("Subset name, CG_F1, IL_MCC, pmF1")
    print(results)


if __name__ == "__main__":
    main()