|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import warnings |
|
|
|
|
|
warnings.filterwarnings("ignore", category=RuntimeWarning) |
|
|
import os.path as osp |
|
|
import time |
|
|
import argparse |
|
|
import json |
|
|
import numpy as np |
|
|
import multiprocessing as mp |
|
|
import pandas as pd |
|
|
from pycocotools import mask as cocomask |
|
|
|
|
|
from third_parts.revos.utils.metircs import db_eval_iou, db_eval_boundary, get_r2vos_accuracy, get_r2vos_robustness |
|
|
|
|
|
NUM_WOEKERS = 128 |
|
|
|
|
|
def eval_queue(q, rank, out_dict): |
|
|
while not q.empty(): |
|
|
vid_name, exp = q.get() |
|
|
vid = exp_dict[vid_name] |
|
|
exp_name = f'{vid_name}_{exp}' |
|
|
|
|
|
pred = mask_pred_dict[vid_name][exp] |
|
|
|
|
|
vid_len, h, w = len(vid['frames']), vid['height'], vid['width'] |
|
|
gt_masks = np.zeros((vid_len, h, w), dtype=np.uint8) |
|
|
pred_masks = np.zeros((vid_len, h, w), dtype=np.uint8) |
|
|
|
|
|
anno_ids = vid['expressions'][exp]['anno_id'] |
|
|
|
|
|
for frame_idx, frame_name in enumerate(vid['frames']): |
|
|
|
|
|
for anno_id in anno_ids: |
|
|
mask_rle = mask_dict[str(anno_id)][frame_idx] |
|
|
if mask_rle: |
|
|
gt_masks[frame_idx] += cocomask.decode(mask_rle) |
|
|
|
|
|
|
|
|
pred_mask = cocomask.decode(pred['prediction_masks'][frame_idx]) |
|
|
pred_masks[frame_idx] += pred_mask |
|
|
j = db_eval_iou(gt_masks, pred_masks).mean() |
|
|
f = db_eval_boundary(gt_masks, pred_masks).mean() |
|
|
a = get_r2vos_accuracy(gt_masks, pred_masks).mean() |
|
|
|
|
|
out_dict[exp_name] = [j, f, a] |
|
|
|
|
|
|
|
|
if __name__ == '__main__': |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument("pred_path", type=str, ) |
|
|
parser.add_argument("--exp_path", type=str, default="data/video_datas/revos/meta_expressions_valid_.json") |
|
|
parser.add_argument("--mask_path", type=str, default="data/video_datas/revos/mask_dict.json") |
|
|
parser.add_argument("--save_json_name", type=str, default="revos_valid.json") |
|
|
parser.add_argument("--save_csv_name", type=str, default="revos_valid.csv") |
|
|
args = parser.parse_args() |
|
|
queue = mp.Queue() |
|
|
exp_dict = json.load(open(args.exp_path))['videos'] |
|
|
mask_dict = json.load(open(args.mask_path)) |
|
|
|
|
|
mask_pred = json.load(open(args.pred_path)) |
|
|
|
|
|
shared_exp_dict = mp.Manager().dict(exp_dict) |
|
|
shared_mask_dict = mp.Manager().dict(mask_dict) |
|
|
output_dict = mp.Manager().dict() |
|
|
|
|
|
mask_pred_dict = mp.Manager().dict(mask_pred) |
|
|
|
|
|
|
|
|
for idx, vid_name in enumerate(exp_dict): |
|
|
vid = exp_dict[vid_name] |
|
|
for exp in vid['expressions']: |
|
|
queue.put([vid_name, exp]) |
|
|
|
|
|
start_time = time.time() |
|
|
if NUM_WOEKERS > 1: |
|
|
processes = [] |
|
|
for rank in range(NUM_WOEKERS): |
|
|
p = mp.Process(target=eval_queue, args=(queue, rank, output_dict)) |
|
|
p.start() |
|
|
processes.append(p) |
|
|
|
|
|
for p in processes: |
|
|
p.join() |
|
|
else: |
|
|
eval_queue(queue, 0, output_dict) |
|
|
|
|
|
|
|
|
output_json_path = osp.join(osp.dirname(args.pred_path), args.save_json_name) |
|
|
output_csv_path = osp.join(osp.dirname(args.pred_path), args.save_csv_name) |
|
|
|
|
|
data_list = [] |
|
|
for videxp, (j, f, a) in output_dict.items(): |
|
|
vid_name, exp = videxp.rsplit('_', 1) |
|
|
data = {} |
|
|
|
|
|
data['video_name'] = vid_name |
|
|
data['exp_id'] = exp |
|
|
data['exp'] = exp_dict[vid_name]['expressions'][exp]['exp'] |
|
|
data['videxp'] = videxp |
|
|
data['J'] = round(100 * j, 2) |
|
|
data['F'] = round(100 * f, 2) |
|
|
data['JF'] = round(100 * (j + f) / 2, 2) |
|
|
data['A'] = round(100 * a, 2) |
|
|
data['type_id'] = exp_dict[vid_name]['expressions'][exp]['type_id'] |
|
|
|
|
|
data_list.append(data) |
|
|
|
|
|
is_long = lambda x: x['type_id'] == 0 |
|
|
is_short = lambda x: x['type_id'] == 1 |
|
|
|
|
|
j_referring = np.array([d['J'] for d in data_list if is_long(d)]).mean() |
|
|
f_referring = np.array([d['F'] for d in data_list if is_long(d)]).mean() |
|
|
a_referring = np.array([d['A'] for d in data_list if is_long(d)]).mean() |
|
|
jf_referring = (j_referring + f_referring) / 2 |
|
|
|
|
|
j_reason = np.array([d['J'] for d in data_list if is_short(d)]).mean() |
|
|
f_reason = np.array([d['F'] for d in data_list if is_short(d)]).mean() |
|
|
a_reason = np.array([d['A'] for d in data_list if is_short(d)]).mean() |
|
|
jf_reason = (j_reason + f_reason) / 2 |
|
|
|
|
|
j_referring_reason = (j_referring + j_reason) / 2 |
|
|
f_referring_reason = (f_referring + f_reason) / 2 |
|
|
a_referring_reason = (a_referring + a_reason) / 2 |
|
|
jf_referring_reason = (jf_referring + jf_reason) / 2 |
|
|
|
|
|
results = { |
|
|
"long": { |
|
|
"J" : j_referring, |
|
|
"F" : f_referring, |
|
|
"A" : a_referring, |
|
|
"JF": jf_referring |
|
|
}, |
|
|
"short": { |
|
|
"J" : j_reason, |
|
|
"F" : f_reason, |
|
|
"A" : a_reason, |
|
|
"JF": jf_reason |
|
|
}, |
|
|
"overall": { |
|
|
"J" : j_referring_reason, |
|
|
"F" : f_referring_reason, |
|
|
"A" : a_referring_reason, |
|
|
"JF": jf_referring_reason |
|
|
} |
|
|
} |
|
|
|
|
|
print(results) |
|
|
with open(output_json_path, 'w') as f: |
|
|
json.dump(results, f, indent=4) |
|
|
print(f"Results saved to {output_json_path}") |
|
|
|
|
|
data4csv = {} |
|
|
for data in data_list: |
|
|
for k, v in data.items(): |
|
|
data4csv[k] = data4csv.get(k, []) + [v] |
|
|
|
|
|
df = pd.DataFrame(data4csv) |
|
|
df.to_csv(output_csv_path, index=False) |
|
|
print(f"Results saved to {output_csv_path}") |
|
|
|
|
|
end_time = time.time() |
|
|
total_time = end_time - start_time |
|
|
print("time: %.4f s" %(total_time)) |
|
|
|