|
|
import numpy as np |
|
|
from .SODA.soda import SODA |
|
|
from .SODA.dataset import ANETCaptions |
|
|
|
|
|
def eval_tool(prediction, referneces=None, metric='Meteor', soda_type='c', verbose=False, print_matrix=False): |
|
|
|
|
|
args = type('args', (object,), {})() |
|
|
args.prediction = prediction |
|
|
args.references = referneces |
|
|
args.metric = metric |
|
|
args.soda_type = soda_type |
|
|
args.tious = [0.3, 0.5, 0.7, 0.9] |
|
|
args.verbose = verbose |
|
|
args.multi_reference = False |
|
|
|
|
|
data = ANETCaptions.from_load_files(args.references, |
|
|
args.prediction, |
|
|
multi_reference=args.multi_reference, |
|
|
verbose=args.verbose, |
|
|
) |
|
|
data.preprocess() |
|
|
if args.soda_type == 'a': |
|
|
tious = args.tious |
|
|
else: |
|
|
tious = None |
|
|
evaluator = SODA(data, |
|
|
soda_type=args.soda_type, |
|
|
tious=tious, |
|
|
scorer=args.metric, |
|
|
verbose=args.verbose, |
|
|
print_matrix=print_matrix |
|
|
) |
|
|
result = evaluator.evaluate() |
|
|
|
|
|
return result |
|
|
|
|
|
def eval_soda(p, ref_list,verbose=False, print_matrix=False): |
|
|
score_sum = [] |
|
|
for ref in ref_list: |
|
|
r = eval_tool(prediction=p, referneces=[ref], verbose=verbose, soda_type='c', print_matrix=print_matrix) |
|
|
score_sum.append(r['Meteor']) |
|
|
soda_avg = np.mean(score_sum, axis=0) |
|
|
soda_c_avg = soda_avg[-1] |
|
|
results = {'soda_c': soda_c_avg} |
|
|
return results |