FangSen9000
Attempted to submit 4 changes, although the reasoning degraded, the reasoning could still run.
1eb306c
# coding: utf-8
"""
This module holds various MT evaluation metrics.
"""
import argparse
import sacrebleu
import mscoco_rouge
import numpy as np
import phoenix_cleanup as phoenix_utils
WER_COST_DEL = 3
WER_COST_INS = 3
WER_COST_SUB = 4
def chrf(references, hypotheses):
"""
Character F-score from sacrebleu
:param hypotheses: list of hypotheses (strings)
:param references: list of references (strings)
:return:
"""
return (
sacrebleu.corpus_chrf(hypotheses=hypotheses, references=references).score * 100
)
def bleu(references, hypotheses):
"""
Raw corpus BLEU from sacrebleu (without tokenization)
:param hypotheses: list of hypotheses (strings)
:param references: list of references (strings)
:return:
"""
bleu_scores = sacrebleu.raw_corpus_bleu(
sys_stream=hypotheses, ref_streams=[references]
).scores
scores = {}
for n in range(len(bleu_scores)):
scores["bleu" + str(n + 1)] = bleu_scores[n]
return scores
def sableu(references, hypotheses, tokenizer):
"""
Sacrebleu (with tokenization)
:param hypotheses: list of hypotheses (strings)
:param references: list of references (strings)
:return:
"""
bleu_scores = sacrebleu.corpus_bleu(
sys_stream=hypotheses, ref_streams=[references], tokenize=tokenizer,
).scores
scores = {}
for n in range(len(bleu_scores)):
scores["bleu" + str(n + 1)] = bleu_scores[n]
return scores
def token_accuracy(references, hypotheses, level="word"):
"""
Compute the accuracy of hypothesis tokens: correct tokens / all tokens
Tokens are correct if they appear in the same position in the reference.
:param hypotheses: list of hypotheses (strings)
:param references: list of references (strings)
:param level: segmentation level, either "word", "bpe", or "char"
:return:
"""
correct_tokens = 0
all_tokens = 0
split_char = " " if level in ["word", "bpe"] else ""
assert len(hypotheses) == len(references)
for hyp, ref in zip(hypotheses, references):
all_tokens += len(hyp)
for h_i, r_i in zip(hyp.split(split_char), ref.split(split_char)):
# min(len(h), len(r)) tokens considered
if h_i == r_i:
correct_tokens += 1
return (correct_tokens / all_tokens) * 100 if all_tokens > 0 else 0.0
def sequence_accuracy(references, hypotheses):
"""
Compute the accuracy of hypothesis tokens: correct tokens / all tokens
Tokens are correct if they appear in the same position in the reference.
:param hypotheses: list of hypotheses (strings)
:param references: list of references (strings)
:return:
"""
assert len(hypotheses) == len(references)
correct_sequences = sum(
[1 for (hyp, ref) in zip(hypotheses, references) if hyp == ref]
)
return (correct_sequences / len(hypotheses)) * 100 if hypotheses else 0.0
def rouge(references, hypotheses):
rouge_score = 0
n_seq = len(hypotheses)
for h, r in zip(hypotheses, references):
rouge_score += mscoco_rouge.calc_score(hypotheses=[h], references=[r]) / n_seq
return rouge_score * 100
def wer_list(references, hypotheses):
total_error = total_del = total_ins = total_sub = total_ref_len = 0
for r, h in zip(references, hypotheses):
res = wer_single(r=r, h=h)
total_error += res["num_err"]
total_del += res["num_del"]
total_ins += res["num_ins"]
total_sub += res["num_sub"]
total_ref_len += res["num_ref"]
wer = (total_error / total_ref_len) * 100
del_rate = (total_del / total_ref_len) * 100
ins_rate = (total_ins / total_ref_len) * 100
sub_rate = (total_sub / total_ref_len) * 100
return {
"wer": wer,
"del_rate": del_rate,
"ins_rate": ins_rate,
"sub_rate": sub_rate,
}
def wer_single(r, h):
r = r.strip().split()
h = h.strip().split()
edit_distance_matrix = edit_distance(r=r, h=h)
alignment, alignment_out = get_alignment(r=r, h=h, d=edit_distance_matrix)
num_cor = np.sum([s == "C" for s in alignment])
num_del = np.sum([s == "D" for s in alignment])
num_ins = np.sum([s == "I" for s in alignment])
num_sub = np.sum([s == "S" for s in alignment])
num_err = num_del + num_ins + num_sub
num_ref = len(r)
return {
"alignment": alignment,
"alignment_out": alignment_out,
"num_cor": num_cor,
"num_del": num_del,
"num_ins": num_ins,
"num_sub": num_sub,
"num_err": num_err,
"num_ref": num_ref,
}
def edit_distance(r, h):
"""
Original Code from https://github.com/zszyellow/WER-in-python/blob/master/wer.py
This function is to calculate the edit distance of reference sentence and the hypothesis sentence.
Main algorithm used is dynamic programming.
Attributes:
r -> the list of words produced by splitting reference sentence.
h -> the list of words produced by splitting hypothesis sentence.
"""
d = np.zeros((len(r) + 1) * (len(h) + 1), dtype=np.uint8).reshape(
(len(r) + 1, len(h) + 1)
)
for i in range(len(r) + 1):
for j in range(len(h) + 1):
if i == 0:
# d[0][j] = j
d[0][j] = j * WER_COST_INS
elif j == 0:
d[i][0] = i * WER_COST_DEL
for i in range(1, len(r) + 1):
for j in range(1, len(h) + 1):
if r[i - 1] == h[j - 1]:
d[i][j] = d[i - 1][j - 1]
else:
substitute = d[i - 1][j - 1] + WER_COST_SUB
insert = d[i][j - 1] + WER_COST_INS
delete = d[i - 1][j] + WER_COST_DEL
d[i][j] = min(substitute, insert, delete)
return d
def get_alignment(r, h, d):
"""
Original Code from https://github.com/zszyellow/WER-in-python/blob/master/wer.py
This function is to get the list of steps in the process of dynamic programming.
Attributes:
r -> the list of words produced by splitting reference sentence.
h -> the list of words produced by splitting hypothesis sentence.
d -> the matrix built when calculating the editing distance of h and r.
"""
x = len(r)
y = len(h)
max_len = 3 * (x + y)
alignlist = []
align_ref = ""
align_hyp = ""
alignment = ""
while True:
if (x <= 0 and y <= 0) or (len(alignlist) > max_len):
break
elif x >= 1 and y >= 1 and d[x][y] == d[x - 1][y - 1] and r[x - 1] == h[y - 1]:
align_hyp = " " + h[y - 1] + align_hyp
align_ref = " " + r[x - 1] + align_ref
alignment = " " * (len(r[x - 1]) + 1) + alignment
alignlist.append("C")
x = max(x - 1, 0)
y = max(y - 1, 0)
elif x >= 1 and y >= 1 and d[x][y] == d[x - 1][y - 1] + WER_COST_SUB:
ml = max(len(h[y - 1]), len(r[x - 1]))
align_hyp = " " + h[y - 1].ljust(ml) + align_hyp
align_ref = " " + r[x - 1].ljust(ml) + align_ref
alignment = " " + "S" + " " * (ml - 1) + alignment
alignlist.append("S")
x = max(x - 1, 0)
y = max(y - 1, 0)
elif y >= 1 and d[x][y] == d[x][y - 1] + WER_COST_INS:
align_hyp = " " + h[y - 1] + align_hyp
align_ref = " " + "*" * len(h[y - 1]) + align_ref
alignment = " " + "I" + " " * (len(h[y - 1]) - 1) + alignment
alignlist.append("I")
x = max(x, 0)
y = max(y - 1, 0)
else:
align_hyp = " " + "*" * len(r[x - 1]) + align_hyp
align_ref = " " + r[x - 1] + align_ref
alignment = " " + "D" + " " * (len(r[x - 1]) - 1) + alignment
alignlist.append("D")
x = max(x - 1, 0)
y = max(y, 0)
align_ref = align_ref[1:]
align_hyp = align_hyp[1:]
alignment = alignment[1:]
return (
alignlist[::-1],
{"align_ref": align_ref, "align_hyp": align_hyp, "alignment": alignment},
)
if __name__ == "__main__":
arg_parser = argparse.ArgumentParser(
description="SLTUnet Evaluator: quality evaluation for sign language translation",
formatter_class=argparse.RawDescriptionHelpFormatter,
)
arg_parser.add_argument(
"--task",
"-t",
choices=["slt", "slr"],
type=str,
default=None,
required=True,
help="the task for evaluation, either sign language translation (slt) or sign langauge recognition (slr)",
)
arg_parser.add_argument(
"--hypothesis",
"-hyp",
type=str,
default=None,
required=True,
help="Model output or system generation.",
)
arg_parser.add_argument(
"--reference",
"-ref",
type=str,
default=None,
required=True,
help="Gold reference",
)
arg_parser.add_argument(
"--tokenize",
"-tok",
choices=sacrebleu.TOKENIZERS.keys(),
default="13a",
help="tokenization method to use",
)
arg_parser.add_argument(
"--phoenix",
default=False,
action="store_true",
help="Perform evaluation for Phoenix 2014T (special preprocessing will be applied to glosses)",
)
args = arg_parser.parse_args()
references = [l.strip() for l in open(args.reference, 'r')]
hypotheses = [l.strip() for l in open(args.hypothesis, 'r')]
if args.task == "slr": # sign language recognition requires WER
if args.phoenix:
references = [phoenix_utils.clean_phoenix_2014_trans(r) for r in references]
hypotheses = [phoenix_utils.clean_phoenix_2014_trans(h) for h in hypotheses]
print('Wer', wer_list(references, hypotheses))
else:
if args.tokenize == "none": # default result
print('BLEU', bleu(references, hypotheses))
print('Rouge', rouge(references, hypotheses))
else: # sacrebleu
print('Signature: BLEU+case.mixed+numrefs.1+smooth.exp+tok.%s+version.1.4.2' % args.tokenize)
print('BLEU', sableu(references, hypotheses, args.tokenize))
print('Signature: chrF2+case.mixed+numchars.6+numrefs.1+space.False+version.1.4.2')
print('Chrf', chrf(references, hypotheses))