# coding: utf-8 """ This module holds various MT evaluation metrics. """ import argparse import sacrebleu import mscoco_rouge import numpy as np import phoenix_cleanup as phoenix_utils WER_COST_DEL = 3 WER_COST_INS = 3 WER_COST_SUB = 4 def chrf(references, hypotheses): """ Character F-score from sacrebleu :param hypotheses: list of hypotheses (strings) :param references: list of references (strings) :return: """ return ( sacrebleu.corpus_chrf(hypotheses=hypotheses, references=references).score * 100 ) def bleu(references, hypotheses): """ Raw corpus BLEU from sacrebleu (without tokenization) :param hypotheses: list of hypotheses (strings) :param references: list of references (strings) :return: """ bleu_scores = sacrebleu.raw_corpus_bleu( sys_stream=hypotheses, ref_streams=[references] ).scores scores = {} for n in range(len(bleu_scores)): scores["bleu" + str(n + 1)] = bleu_scores[n] return scores def sableu(references, hypotheses, tokenizer): """ Sacrebleu (with tokenization) :param hypotheses: list of hypotheses (strings) :param references: list of references (strings) :return: """ bleu_scores = sacrebleu.corpus_bleu( sys_stream=hypotheses, ref_streams=[references], tokenize=tokenizer, ).scores scores = {} for n in range(len(bleu_scores)): scores["bleu" + str(n + 1)] = bleu_scores[n] return scores def token_accuracy(references, hypotheses, level="word"): """ Compute the accuracy of hypothesis tokens: correct tokens / all tokens Tokens are correct if they appear in the same position in the reference. :param hypotheses: list of hypotheses (strings) :param references: list of references (strings) :param level: segmentation level, either "word", "bpe", or "char" :return: """ correct_tokens = 0 all_tokens = 0 split_char = " " if level in ["word", "bpe"] else "" assert len(hypotheses) == len(references) for hyp, ref in zip(hypotheses, references): all_tokens += len(hyp) for h_i, r_i in zip(hyp.split(split_char), ref.split(split_char)): # min(len(h), len(r)) tokens considered if h_i == r_i: correct_tokens += 1 return (correct_tokens / all_tokens) * 100 if all_tokens > 0 else 0.0 def sequence_accuracy(references, hypotheses): """ Compute the accuracy of hypothesis tokens: correct tokens / all tokens Tokens are correct if they appear in the same position in the reference. :param hypotheses: list of hypotheses (strings) :param references: list of references (strings) :return: """ assert len(hypotheses) == len(references) correct_sequences = sum( [1 for (hyp, ref) in zip(hypotheses, references) if hyp == ref] ) return (correct_sequences / len(hypotheses)) * 100 if hypotheses else 0.0 def rouge(references, hypotheses): rouge_score = 0 n_seq = len(hypotheses) for h, r in zip(hypotheses, references): rouge_score += mscoco_rouge.calc_score(hypotheses=[h], references=[r]) / n_seq return rouge_score * 100 def wer_list(references, hypotheses): total_error = total_del = total_ins = total_sub = total_ref_len = 0 for r, h in zip(references, hypotheses): res = wer_single(r=r, h=h) total_error += res["num_err"] total_del += res["num_del"] total_ins += res["num_ins"] total_sub += res["num_sub"] total_ref_len += res["num_ref"] wer = (total_error / total_ref_len) * 100 del_rate = (total_del / total_ref_len) * 100 ins_rate = (total_ins / total_ref_len) * 100 sub_rate = (total_sub / total_ref_len) * 100 return { "wer": wer, "del_rate": del_rate, "ins_rate": ins_rate, "sub_rate": sub_rate, } def wer_single(r, h): r = r.strip().split() h = h.strip().split() edit_distance_matrix = edit_distance(r=r, h=h) alignment, alignment_out = get_alignment(r=r, h=h, d=edit_distance_matrix) num_cor = np.sum([s == "C" for s in alignment]) num_del = np.sum([s == "D" for s in alignment]) num_ins = np.sum([s == "I" for s in alignment]) num_sub = np.sum([s == "S" for s in alignment]) num_err = num_del + num_ins + num_sub num_ref = len(r) return { "alignment": alignment, "alignment_out": alignment_out, "num_cor": num_cor, "num_del": num_del, "num_ins": num_ins, "num_sub": num_sub, "num_err": num_err, "num_ref": num_ref, } def edit_distance(r, h): """ Original Code from https://github.com/zszyellow/WER-in-python/blob/master/wer.py This function is to calculate the edit distance of reference sentence and the hypothesis sentence. Main algorithm used is dynamic programming. Attributes: r -> the list of words produced by splitting reference sentence. h -> the list of words produced by splitting hypothesis sentence. """ d = np.zeros((len(r) + 1) * (len(h) + 1), dtype=np.uint8).reshape( (len(r) + 1, len(h) + 1) ) for i in range(len(r) + 1): for j in range(len(h) + 1): if i == 0: # d[0][j] = j d[0][j] = j * WER_COST_INS elif j == 0: d[i][0] = i * WER_COST_DEL for i in range(1, len(r) + 1): for j in range(1, len(h) + 1): if r[i - 1] == h[j - 1]: d[i][j] = d[i - 1][j - 1] else: substitute = d[i - 1][j - 1] + WER_COST_SUB insert = d[i][j - 1] + WER_COST_INS delete = d[i - 1][j] + WER_COST_DEL d[i][j] = min(substitute, insert, delete) return d def get_alignment(r, h, d): """ Original Code from https://github.com/zszyellow/WER-in-python/blob/master/wer.py This function is to get the list of steps in the process of dynamic programming. Attributes: r -> the list of words produced by splitting reference sentence. h -> the list of words produced by splitting hypothesis sentence. d -> the matrix built when calculating the editing distance of h and r. """ x = len(r) y = len(h) max_len = 3 * (x + y) alignlist = [] align_ref = "" align_hyp = "" alignment = "" while True: if (x <= 0 and y <= 0) or (len(alignlist) > max_len): break elif x >= 1 and y >= 1 and d[x][y] == d[x - 1][y - 1] and r[x - 1] == h[y - 1]: align_hyp = " " + h[y - 1] + align_hyp align_ref = " " + r[x - 1] + align_ref alignment = " " * (len(r[x - 1]) + 1) + alignment alignlist.append("C") x = max(x - 1, 0) y = max(y - 1, 0) elif x >= 1 and y >= 1 and d[x][y] == d[x - 1][y - 1] + WER_COST_SUB: ml = max(len(h[y - 1]), len(r[x - 1])) align_hyp = " " + h[y - 1].ljust(ml) + align_hyp align_ref = " " + r[x - 1].ljust(ml) + align_ref alignment = " " + "S" + " " * (ml - 1) + alignment alignlist.append("S") x = max(x - 1, 0) y = max(y - 1, 0) elif y >= 1 and d[x][y] == d[x][y - 1] + WER_COST_INS: align_hyp = " " + h[y - 1] + align_hyp align_ref = " " + "*" * len(h[y - 1]) + align_ref alignment = " " + "I" + " " * (len(h[y - 1]) - 1) + alignment alignlist.append("I") x = max(x, 0) y = max(y - 1, 0) else: align_hyp = " " + "*" * len(r[x - 1]) + align_hyp align_ref = " " + r[x - 1] + align_ref alignment = " " + "D" + " " * (len(r[x - 1]) - 1) + alignment alignlist.append("D") x = max(x - 1, 0) y = max(y, 0) align_ref = align_ref[1:] align_hyp = align_hyp[1:] alignment = alignment[1:] return ( alignlist[::-1], {"align_ref": align_ref, "align_hyp": align_hyp, "alignment": alignment}, ) if __name__ == "__main__": arg_parser = argparse.ArgumentParser( description="SLTUnet Evaluator: quality evaluation for sign language translation", formatter_class=argparse.RawDescriptionHelpFormatter, ) arg_parser.add_argument( "--task", "-t", choices=["slt", "slr"], type=str, default=None, required=True, help="the task for evaluation, either sign language translation (slt) or sign langauge recognition (slr)", ) arg_parser.add_argument( "--hypothesis", "-hyp", type=str, default=None, required=True, help="Model output or system generation.", ) arg_parser.add_argument( "--reference", "-ref", type=str, default=None, required=True, help="Gold reference", ) arg_parser.add_argument( "--tokenize", "-tok", choices=sacrebleu.TOKENIZERS.keys(), default="13a", help="tokenization method to use", ) arg_parser.add_argument( "--phoenix", default=False, action="store_true", help="Perform evaluation for Phoenix 2014T (special preprocessing will be applied to glosses)", ) args = arg_parser.parse_args() references = [l.strip() for l in open(args.reference, 'r')] hypotheses = [l.strip() for l in open(args.hypothesis, 'r')] if args.task == "slr": # sign language recognition requires WER if args.phoenix: references = [phoenix_utils.clean_phoenix_2014_trans(r) for r in references] hypotheses = [phoenix_utils.clean_phoenix_2014_trans(h) for h in hypotheses] print('Wer', wer_list(references, hypotheses)) else: if args.tokenize == "none": # default result print('BLEU', bleu(references, hypotheses)) print('Rouge', rouge(references, hypotheses)) else: # sacrebleu print('Signature: BLEU+case.mixed+numrefs.1+smooth.exp+tok.%s+version.1.4.2' % args.tokenize) print('BLEU', sableu(references, hypotheses, args.tokenize)) print('Signature: chrF2+case.mixed+numchars.6+numrefs.1+space.False+version.1.4.2') print('Chrf', chrf(references, hypotheses))