simplecloud's picture
Upload folder using huggingface_hub
fca4fc0 verified
# --------------------------------------------------------
# evaluation scripts for dense video captioning, support python 3
# Modified from https://github.com/ranjaykrishna/densevid_eval/tree/9d4045aced3d827834a5d2da3c9f0692e3f33c1c
# --------------------------------------------------------
# Dense-Captioning Events in Videos Eval
# Copyright (c) 2017 Ranjay Krishna
# Licensed under The MIT License [see LICENSE for details]
# Written by Ranjay Krishna
# --------------------------------------------------------
import argparse
import json
import random
import string
import sys
import time
# sys.path.insert(0, './coco-caption') # Hack to allow the import of pycocoeval
from pycocoevalcap.tokenizer.ptbtokenizer import PTBTokenizer
from pycocoevalcap.meteor.meteor import Meteor
Set = set
import numpy as np
def random_string(string_length):
letters = string.ascii_lowercase
return ''.join(random.choice(letters) for i in range(string_length))
def remove_nonascii(text):
return ''.join([i if ord(i) < 128 else ' ' for i in text])
class ANETcaptions(object):
PREDICTION_FIELDS = ['results', 'version', 'external_data']
def __init__(self, ground_truth_filenames=None, prediction_filename=None,
tious=None, distances=[1, 3, 5, 10, 30, 60], max_proposals=1000,
prediction_fields=PREDICTION_FIELDS, verbose=False, no_lang_eval=False):
# Check that the gt and submission files exist and load them
if len(tious) == 0:
raise IOError('Please input a valid tIoU.')
if not ground_truth_filenames:
raise IOError('Please input a valid ground truth file.')
if not prediction_filename:
raise IOError('Please input a valid prediction file.')
self.verbose = verbose
self.no_lang_eval = no_lang_eval
self.tious = tious
self.distances = distances
self.max_proposals = max_proposals
self.pred_fields = prediction_fields
self.ground_truths = self.import_ground_truths(ground_truth_filenames)
self.prediction = self.import_prediction(prediction_filename)
self.ground_truths_keys = [vid for gt in self.ground_truths for vid in gt]
print('available video number', len(set(self.ground_truths_keys) & set(self.prediction.keys())))
# Set up scorers
if not self.no_lang_eval:
self.tokenizer = PTBTokenizer()
self.scorers = [
(Meteor(), "METEOR"),
]
def import_prediction(self, prediction_filename):
if self.verbose:
print("| Loading submission...")
if isinstance(prediction_filename, dict):
submission = prediction_filename
else:
submission = json.load(open(prediction_filename))
# if not all([field in submission.keys() for field in self.pred_fields]):
# raise IOError('Please input a valid ground truth file.')
# Ensure that every video is limited to the correct maximum number of proposals.
results = {}
for vid_id in submission['results']:
results[vid_id] = submission['results'][vid_id][:self.max_proposals]
return results
def import_ground_truths(self, filenames):
gts = []
self.n_ref_vids = Set()
for filename in filenames:
if isinstance(filename, dict):
gt = filename
else:
gt = json.load(open(filename))
self.n_ref_vids.update(gt.keys())
gts.append(gt)
if self.verbose:
print("| Loading GT. #files: %d, #videos: %d" % (len(filenames), len(self.n_ref_vids)))
return gts
def iou(self, interval_1, interval_2):
start_i, end_i = interval_1[0], interval_1[1]
start, end = interval_2[0], interval_2[1]
intersection = max(0, min(end, end_i) - max(start, start_i))
union = min(max(end, end_i) - min(start, start_i), end - start + end_i - start_i)
iou = float(intersection) / (union + 1e-8)
return iou
def check_gt_exists(self, vid_id):
for gt in self.ground_truths:
if vid_id in gt:
return True
return False
def get_gt_vid_ids(self):
vid_ids = set([])
for gt in self.ground_truths:
vid_ids |= set(gt.keys())
return list(vid_ids)
def evaluate(self):
aggregator = {}
self.scores = {}
if not self.no_lang_eval:
for tiou in self.tious:
scores = self.evaluate_tiou(tiou)
for metric, score in scores.items():
if metric not in self.scores:
self.scores[metric] = []
self.scores[metric].append(score)
if True:
# if self.verbose:
self.scores['Recall'] = []
self.scores['Precision'] = []
self.scores['F1'] = []
for tiou in self.tious:
precision, recall = self.evaluate_detection(tiou)
self.scores['Recall'].append(recall)
self.scores['Precision'].append(precision)
self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.)
for tiou in self.distances:
precision, recall = self.evaluate_navigation(tiou)
self.scores['Recall'].append(recall)
self.scores['Precision'].append(precision)
self.scores['F1'].append(2 * recall * precision / (recall + precision) if recall + precision else 0.)
def evaluate_detection(self, tiou):
gt_vid_ids = self.get_gt_vid_ids()
# Recall is the percentage of ground truth that is covered by the predictions
# Precision is the percentage of predictions that are valid
recall = []
precision = []
for vid_i, vid_id in enumerate(gt_vid_ids):
if vid_id not in self.prediction: # missing video
continue
best_recall = 0
best_precision = 0
for gt in self.ground_truths:
if vid_id not in gt:
continue
refs = gt[vid_id]
ref_set_covered = set([])
pred_set_covered = set([])
num_gt = 0
num_pred = 0
if vid_id in self.prediction:
for pred_i, pred in enumerate(self.prediction[vid_id]):
pred_timestamp = pred['timestamp']
for ref_i, ref_timestamp in enumerate(refs['timestamps']):
if self.iou(pred_timestamp, ref_timestamp) > tiou:
ref_set_covered.add(ref_i)
pred_set_covered.add(pred_i)
new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1)
best_precision = max(best_precision, new_precision)
new_recall = float(len(ref_set_covered)) / len(refs['timestamps'])
best_recall = max(best_recall, new_recall)
recall.append(best_recall)
precision.append(best_precision)
return sum(precision) / len(precision), sum(recall) / len(recall)
def evaluate_navigation(self, tiou):
gt_vid_ids = self.get_gt_vid_ids()
# Recall is the percentage of ground truth that is covered by the predictions
# Precision is the percentage of predictions that are valid
recall = []
precision = []
for vid_i, vid_id in enumerate(gt_vid_ids):
if vid_id not in self.prediction: # missing video
continue
best_recall = 0
best_precision = 0
for gt in self.ground_truths:
if vid_id not in gt:
continue
refs = gt[vid_id]
ref_set_covered = set([])
pred_set_covered = set([])
num_gt = 0
num_pred = 0
if vid_id in self.prediction:
for pred_i, pred in enumerate(self.prediction[vid_id]):
pred_timestamp = pred['timestamp']
for ref_i, ref_timestamp in enumerate(refs['timestamps']):
if abs(pred_timestamp[0] - ref_timestamp[0]) < tiou:
ref_set_covered.add(ref_i)
pred_set_covered.add(pred_i)
new_precision = float(len(pred_set_covered)) / max(len(self.prediction[vid_id]), 1)
best_precision = max(best_precision, new_precision)
new_recall = float(len(ref_set_covered)) / len(refs['timestamps'])
best_recall = max(best_recall, new_recall)
recall.append(best_recall)
precision.append(best_precision)
return sum(precision) / len(precision), sum(recall) / len(recall)
def evaluate_tiou(self, tiou):
# This method averages the tIoU precision from METEOR, Bleu, etc. across videos
res = {}
gts = {}
gt_vid_ids = self.get_gt_vid_ids()
unique_index = 0
# video id to unique caption ids mapping
vid2capid = {}
cur_res = {}
cur_gts = {}
for vid_id in gt_vid_ids:
# If the video does not have a prediction, then we give it no matches
# We set it to empty, and use this as a sanity check later on
if vid_id not in self.prediction: # missing video
continue
# If we do have a prediction, then we find the scores based on all the
# valid tIoU overlaps.
else:
vid2capid[vid_id] = []
# For each prediction, we look at the tIoU with ground truth.
for pred in self.prediction[vid_id]:
has_added = False
for gt in self.ground_truths:
if vid_id not in gt:
continue
gt_captions = gt[vid_id]
for caption_idx, caption_timestamp in enumerate(gt_captions['timestamps']):
if self.iou(pred['timestamp'], caption_timestamp) >= tiou:
cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}]
cur_gts[unique_index] = [
{'caption': remove_nonascii(gt_captions['sentences'][caption_idx])}]
vid2capid[vid_id].append(unique_index)
unique_index += 1
has_added = True
# If the predicted caption does not overlap with any ground truth,
# we should compare it with garbage.
if not has_added:
cur_res[unique_index] = [{'caption': remove_nonascii(pred['sentence'])}]
cur_gts[unique_index] = [{'caption': random_string(random.randint(10, 20))}]
vid2capid[vid_id].append(unique_index)
unique_index += 1
# Each scorer will compute across all videos and take average score
output = {}
for scorer, method in self.scorers:
if self.verbose:
print('computing %s score...' % (scorer.method()))
# For each video, take all the valid pairs (based from tIoU) and compute the score
all_scores = {}
# call tokenizer here for all predictions and gts
tokenize_res = self.tokenizer.tokenize(cur_res)
tokenize_gts = self.tokenizer.tokenize(cur_gts)
# reshape back
for vid in vid2capid.keys():
res[vid] = {index: tokenize_res[index] for index in vid2capid[vid]}
gts[vid] = {index: tokenize_gts[index] for index in vid2capid[vid]}
for vid_id in gt_vid_ids:
if vid_id not in self.prediction: # missing video
continue
if len(res[vid_id]) == 0 or len(gts[vid_id]) == 0:
if type(method) == list:
score = [0] * len(method)
else:
score = 0
else:
score, scores = scorer.compute_score(gts[vid_id], res[vid_id])
all_scores[vid_id] = score
# import ipdb;ipdb.set_trace()
# print(all_scores.values())
if type(method) == list:
scores = np.mean(list(all_scores.values()), axis=0)
for m in range(len(method)):
output[method[m]] = scores[m]
if self.verbose:
print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method[m], output[method[m]]))
else:
output[method] = np.mean(list(all_scores.values()))
if self.verbose:
print("Calculated tIoU: %1.1f, %s: %0.3f" % (tiou, method, output[method]))
return output
def eval_dvc(submission, references, tious=[0.3, 0.5, 0.7, 0.9], distances=[1, 3, 5, 10, 30, 60], max_proposals_per_video=1000, verbose=False, no_lang_eval=False):
# Call coco eval
evaluator = ANETcaptions(ground_truth_filenames=references,
prediction_filename=submission,
tious=tious,
distances=distances,
max_proposals=max_proposals_per_video,
verbose=verbose, no_lang_eval=no_lang_eval)
evaluator.evaluate()
score = evaluator.scores
# print(score)
loc_score = {}
for i, x in enumerate(tious):
for y in ["Recall", "Precision", "F1"]:
loc_score[y + "@" + str(x)] = score[y][i]
for y in ["Recall", "Precision", "F1"]:
loc_score[y] = np.array([score[y][i] for i in range(len(tious))]).mean()
if distances:
for i, x in enumerate(distances):
for y in ["Recall", "Precision", "F1"]:
loc_score[y + "@" + str(x) + "s"] = score[y][len(tious) + i]
avg_eval_score = {key: np.array(value).mean() for key, value in score.items() if key not in ["Recall", "Precision", "F1"]}
avg_eval_score.update(loc_score)
return avg_eval_score
if __name__ == '__main__':
eval_dvc(pred_path, references,
tious=[0.3, 0.5, 0.7, 0.9],
max_proposals_per_video=1000,
verbose=False,
no_lang_eval=False)
eval_soda(pred_path, references, verbose=False)