# -------------------------------------------------------- # evaluation scripts for dense video captioning, support python 3 # Modified from https://github.com/ranjaykrishna/densevid_eval/tree/9d4045aced3d827834a5d2da3c9f0692e3f33c1c # -------------------------------------------------------- # Dense-Captioning Events in Videos Eval # Copyright (c) 2017 Ranjay Krishna # Licensed under The MIT License [see LICENSE for details] # Written by Ranjay Krishna # -------------------------------------------------------- import argparse import json import random import string import sys import time # sys.path.insert(0, './coco-caption') # Hack to allow the import of pycocoeval from pycocoevalcap.tokenizer.ptbtokenizer import PTBTokenizer from pycocoevalcap.meteor.meteor import Meteor from pycocoevalcap.cider.cider import Cider from pycocoevalcap.bleu.bleu import Bleu from pycocoevalcap.rouge.rouge import Rouge Set = set import numpy as np def random_string(string_length): letters = string.ascii_lowercase return ''.join(random.choice(letters) for i in range(string_length)) def remove_nonascii(text): return ''.join([i if ord(i) < 128 else ' ' for i in text]) class ANETcaptions(object): PREDICTION_FIELDS = ['results', 'version', 'external_data'] def __init__(self, ground_truth_filenames=None, prediction_filename=None, tious=None, distances=[1, 3, 5, 10, 30, 60], max_proposals=1000, prediction_fields=PREDICTION_FIELDS, verbose=False, no_lang_eval=False): # Check that the gt and submission files exist and load them if len(tious) == 0: raise IOError('Please input a valid tIoU.') if not ground_truth_filenames: raise IOError('Please input a valid ground truth file.') if not prediction_filename: raise IOError('Please input a valid prediction file.') self.verbose = verbose self.no_lang_eval = no_lang_eval self.tious = tious self.distances = distances self.max_proposals = max_proposals self.pred_fields = prediction_fields self.ground_truths = self.import_ground_truths(ground_truth_filenames) self.prediction = self.import_prediction(prediction_filename) self.ground_truths_keys = [vid for gt in self.ground_truths for vid in gt] print('available video number', len(set(self.ground_truths_keys) & set(self.prediction.keys()))) # Set up scorers if not self.no_lang_eval: self.tokenizer = PTBTokenizer() self.scorers = [ (Meteor(), "METEOR"), (Cider(), "CIDEr"), (Rouge(), "Rouge-L"), (Bleu(4), ["Bleu_1", "Bleu_2", "Bleu_3", "Bleu_4"]), ] def import_prediction(self, prediction_filename): if self.verbose: print("| Loading submission...") if isinstance(prediction_filename, dict): submission = prediction_filename else: submission = json.load(open(prediction_filename)) # if not all([field in submission.keys() for field in self.pred_fields]): # raise IOError('Please input a valid ground truth file.') # Ensure that every video is limited to the correct maximum number of proposals. results = {} for vid_id in submission['results']: results[vid_id] = submission['results'][vid_id][:self.max_proposals] return results def import_ground_truths(self, filenames): gts = [] self.n_ref_vids = Set() for filename in filenames: if isinstance(filename, dict): gt = filename else: gt = json.load(open(filename)) self.n_ref_vids.update(gt.keys()) gts.append(gt) if self.verbose: print("| Loading GT. #files: %d, #videos: %d" % (len(filenames), len(self.n_ref_vids))) return gts def iou(self, interval_1, interval_2): start_i, end_i = interval_1[0], interval_1[1] start, end = interval_2[0], interval_2[1] intersection = max(0, min(end, end_i) - max(start, start_i)) union = min(max(end, end_i) - min(start, start_i), end - start + end_i - start_i) iou = float(intersection) / (union + 1e-8) return iou def check_gt_exists(self, vid_id): for gt in self.ground_truths: if vid_id in gt: return True return False def get_gt_vid_ids(self): vid_ids = set([]) for gt in self.ground_truths: vid_ids |= set(gt.keys()) return list(vid_ids) def evaluate(self): aggregator = {} self.scores = {} if not self.no_lang_eval: for tiou in self.tious: scores = self.evaluate_tiou(tiou) for metric, score in scores.items(): if metric not in self.scores: self.scores[metric] = [] self.scores[metric].append(score) if True: # if self.verbose: self.scores['Recall'] = [] self.scores['Precision'] = [] self.scores['F1'] = [] for tiou in self.tious: precision, recall = self.evaluate_detection(tiou) self.scores['Recall'].append(recall) self.scores['Precision'].append(precision) self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) for tiou in self.distances: precision, recall = self.evaluate_navigation(tiou) self.scores['Recall'].append(recall) self.scores['Precision'].append(precision) self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.) def evaluate_detection(self, tiou): gt_vid_ids = self.get_gt_vid_ids() # Recall is the percentage of ground truth that is covered by the predictions # Precision is the percentage of predictions that are valid recall = [] precision = [] for vid_i, vid_id in enumerate(gt_vid_ids): if vid_id not in self.prediction: # missing video continue best_recall = 0 best_precision = 0 for gt in self.ground_truths: if vid_id not in gt: continue refs = gt[vid_id] ref_set_covered = set([]) pred_set_covered = set([]) num_gt = 0 num_pred = 0 if vid_id in self.prediction: for pred_i, pred in enumerate(self.prediction[vid_id]): pred_timestamp = pred['timestamp'] for ref_i, ref_timestamp in enumerate(refs['timestamps']): if self.iou(pred_timestamp, ref_timestamp) > tiou: ref_set_covered.add(ref_i) pred_set_covered.add(pred_i) new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) best_precision = max(best_precision, new_precision) new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) best_recall = max(best_recall, new_recall) recall.append(best_recall) precision.append(best_precision) return sum(precision) / len(precision), sum(recall) / len(recall) def evaluate_navigation(self, tiou): gt_vid_ids = self.get_gt_vid_ids() # Recall is the percentage of ground truth that is covered by the predictions # Precision is the percentage of predictions that are valid recall = [] precision = [] for vid_i, vid_id in enumerate(gt_vid_ids): if vid_id not in self.prediction: # missing video continue best_recall = 0 best_precision = 0 for gt in self.ground_truths: if vid_id not in gt: continue refs = gt[vid_id] ref_set_covered = set([]) pred_set_covered = set([]) num_gt = 0 num_pred = 0 if vid_id in self.prediction: for pred_i, pred in enumerate(self.prediction[vid_id]): pred_timestamp = pred['timestamp'] for ref_i, ref_timestamp in enumerate(refs['timestamps']): if abs(pred_timestamp[0] - ref_timestamp[0]) < tiou: ref_set_covered.add(ref_i) pred_set_covered.add(pred_i) new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1) best_precision = max(best_precision, new_precision) new_recall = float(len(ref_set_covered)) / len(refs['timestamps']) best_recall = max(best_recall, new_recall) recall.append(best_recall) precision.append(best_precision) return sum(precision) / len(precision), sum(recall) / len(recall) def evaluate_tiou(self, tiou): # This method averages the tIoU precision from METEOR, Bleu, etc. across videos res = {} gts = {} gt_vid_ids = self.get_gt_vid_ids() unique_index = 0 # video id to unique caption ids mapping vid2capid = {} cur_res = {} cur_gts = {} for vid_id in gt_vid_ids: # If the video does not have a prediction, then we give it no matches # We set it to empty, and use this as a sanity check later on if vid_id not in self.prediction: # missing video continue # If we do have a prediction, then we find the scores based on all the # valid tIoU overlaps. else: vid2capid[vid_id] = [] # For each prediction, we look at the tIoU with ground truth. for pred in self.prediction[vid_id]: has_added = False for gt in self.ground_truths: if vid_id not in gt: continue gt_captions = gt[vid_id] for caption_idx, caption_timestamp in enumerate(gt_captions['timestamps']): if self.iou(pred['timestamp'], caption_timestamp) >= tiou: cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] cur_gts[unique_index] = [ {'caption': remove_nonascii(gt_captions['sentences'][caption_idx])}] vid2capid[vid_id].append(unique_index) unique_index += 1 has_added = True # If the predicted caption does not overlap with any ground truth, # we should compare it with garbage. if not has_added: cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}] cur_gts[unique_index] = [{'caption': random_string(random.randint(10, 20))}] vid2capid[vid_id].append(unique_index) unique_index += 1 # Each scorer will compute across all videos and take average score output = {} for scorer, method in self.scorers: if self.verbose: print('computing %s score...' % (scorer.method())) # For each video, take all the valid pairs (based from tIoU) and compute the score all_scores = {} # call tokenizer here for all predictions and gts tokenize_res = self.tokenizer.tokenize(cur_res) tokenize_gts = self.tokenizer.tokenize(cur_gts) # reshape back for vid in vid2capid.keys(): res[vid] = {index: tokenize_res[index] for index in vid2capid[vid]} gts[vid] = {index: tokenize_gts[index] for index in vid2capid[vid]} for vid_id in gt_vid_ids: if vid_id not in self.prediction: # missing video continue if len(res[vid_id]) == 0 or len(gts[vid_id]) == 0: if type(method) == list: score = [0] * len(method) else: score = 0 else: score, scores = scorer.compute_score(gts[vid_id], res[vid_id]) all_scores[vid_id] = score # import ipdb;ipdb.set_trace() # print(all_scores.values()) if type(method) == list: scores = np.mean(list(all_scores.values()), axis=0) for m in range(len(method)): output[method[m]] = scores[m] if self.verbose: print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method[m], output[method[m]])) else: output[method] = np.mean(list(all_scores.values())) if self.verbose: print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method, output[method])) return output def eval_dvc(submission, references, tious=[0.3, 0.5, 0.7, 0.9], distances=[1, 3, 5, 10, 30, 60], max_proposals_per_video=1000, verbose=False, no_lang_eval=False): # Call coco eval evaluator = ANETcaptions(ground_truth_filenames=references, prediction_filename=submission, tious=tious, distances=distances, max_proposals=max_proposals_per_video, verbose=verbose, no_lang_eval=no_lang_eval) evaluator.evaluate() score = evaluator.scores # print(score) loc_score = {} for i, x in enumerate(tious): for y in ["Recall", "Precision", "F1"]: loc_score[y + "@" + str(x)] = score[y][i] for y in ["Recall", "Precision", "F1"]: loc_score[y] = np.array([score[y][i] for i in range(len(tious))]).mean() if distances: for i, x in enumerate(distances): for y in ["Recall", "Precision", "F1"]: loc_score[y + "@" + str(x) + "s"] = score[y][len(tious) + i] avg_eval_score = {key: np.array(value).mean() for key, value in score.items() if key not in ["Recall", "Precision", "F1"]} avg_eval_score.update(loc_score) return avg_eval_score if __name__ == '__main__': eval_dvc(pred_path, references, tious=[0.3, 0.5, 0.7, 0.9], max_proposals_per_video=1000, verbose=False, no_lang_eval=False) eval_soda(pred_path, references, verbose=False)