| |
| |
| |
| |
| |
|
|
| import argparse |
| import json |
| import os |
|
|
| import yaml |
|
|
| from pythainlp import cli |
| from pythainlp.benchmarks import word_tokenization |
| from pythainlp.tools import safe_print |
|
|
|
|
| def _read_file(path): |
| with open(path, "r", encoding="utf-8") as f: |
| lines = map(lambda r: r.strip(), f.readlines()) |
| return list(lines) |
|
|
|
|
| class App: |
| def __init__(self, argv): |
| parser = argparse.ArgumentParser( |
| prog="benchmark", |
| description=( |
| "Benchmark for various tasks;\n" |
| "currently, we have only for word tokenization." |
| ), |
| usage=( |
| "thainlp benchmark [task] [task-options]\n\n" |
| "tasks:\n\n" |
| "word-tokenization benchmark word tokenization\n\n" |
| "--" |
| ), |
| ) |
|
|
| parser.add_argument("task", type=str, help="[word-tokenization]") |
|
|
| args = parser.parse_args(argv[2:3]) |
| cli.exit_if_empty(args.task, parser) |
| task = str.lower(args.task) |
|
|
| task_argv = argv[3:] |
| if task == "word-tokenization": |
| WordTokenizationBenchmark(task, task_argv) |
|
|
|
|
| class WordTokenizationBenchmark: |
| def __init__(self, name, argv): |
| parser = argparse.ArgumentParser(**cli.make_usage("benchmark " + name)) |
|
|
| parser.add_argument( |
| "--input-file", |
| action="store", |
| help="Path to input file to compare against the test file", |
| ) |
|
|
| parser.add_argument( |
| "--test-file", |
| action="store", |
| help="Path to test file i.e. ground truth", |
| ) |
|
|
| parser.add_argument( |
| "--save-details", |
| default=False, |
| action="store_true", |
| help=( |
| "Save comparison details to files (eval-XXX.json" |
| " and eval-details-XXX.json)" |
| ), |
| ) |
|
|
| args = parser.parse_args(argv) |
|
|
| actual = _read_file(args.input_file) |
| expected = _read_file(args.test_file) |
|
|
| assert len(actual) == len( |
| expected |
| ), "Input and test files do not have the same number of samples" |
|
|
| safe_print( |
| "Benchmarking %s against %s with %d samples in total" |
| % (args.input_file, args.test_file, len(actual)) |
| ) |
|
|
| df_raw = word_tokenization.benchmark(expected, actual) |
|
|
| columns = [ |
| "char_level:tp", |
| "char_level:fp", |
| "char_level:tn", |
| "char_level:fn", |
| "word_level:correctly_tokenised_words", |
| "word_level:total_words_in_sample", |
| "word_level:total_words_in_ref_sample", |
| ] |
|
|
| statistics = {} |
|
|
| for c in columns: |
| statistics[c] = float(df_raw[c].sum()) |
|
|
| statistics["char_level:precision"] = statistics["char_level:tp"] / ( |
| statistics["char_level:tp"] + statistics["char_level:fp"] |
| ) |
|
|
| statistics["char_level:recall"] = statistics["char_level:tp"] / ( |
| statistics["char_level:tp"] + statistics["char_level:fn"] |
| ) |
|
|
| statistics["word_level:precision"] = ( |
| statistics["word_level:correctly_tokenised_words"] |
| / statistics["word_level:total_words_in_sample"] |
| ) |
|
|
| statistics["word_level:recall"] = ( |
| statistics["word_level:correctly_tokenised_words"] |
| / statistics["word_level:total_words_in_ref_sample"] |
| ) |
|
|
| safe_print("============== Benchmark Result ==============") |
|
|
| for c in ["tp", "fn", "tn", "fp", "precision", "recall"]: |
| c = f"char_level:{c}" |
| v = statistics[c] |
| safe_print(f"{c:>40s} {v:.4f}") |
|
|
| for c in [ |
| "total_words_in_sample", |
| "total_words_in_ref_sample", |
| "correctly_tokenised_words", |
| "precision", |
| "recall", |
| ]: |
| c = f"word_level:{c}" |
| v = statistics[c] |
| safe_print(f"{c:>40s} {v:.4f}") |
|
|
| if args.save_details: |
| dir_name = os.path.dirname(args.input_file) |
| file_name = args.input_file.split("/")[-1].split(".")[0] |
|
|
| res_path = "%s/eval-%s.yml" % (dir_name, file_name) |
| safe_print("Evaluation result is saved to %s" % res_path) |
|
|
| with open(res_path, "w", encoding="utf-8") as outfile: |
| yaml.dump(statistics, outfile, default_flow_style=False) |
|
|
| res_path = "%s/eval-details-%s.json" % (dir_name, file_name) |
| safe_print("Details of comparisons is saved to %s" % res_path) |
|
|
| with open(res_path, "w", encoding="utf-8") as f: |
| samples = [] |
| for i, r in enumerate(df_raw.to_dict("records")): |
| expected, actual = r["expected"], r["actual"] |
| del r["expected"] |
| del r["actual"] |
|
|
| samples.append( |
| { |
| "metrics": r, |
| "expected": expected, |
| "actual": actual, |
| "id": i, |
| } |
| ) |
|
|
| details = {"metrics": statistics, "samples": samples} |
|
|
| json.dump(details, f, ensure_ascii=False) |
|
|