| | import argparse
|
| |
|
| | from src.utils import read_file, check_data_validity
|
| | from src.layout_evaluation import evaluate_layout
|
| | from src.table_evaluation import evaluate_table
|
| |
|
| |
|
| | def parse_args():
|
| | parser = argparse.ArgumentParser(description="Arguments for evaluation")
|
| | parser.add_argument(
|
| | "--ref_path",
|
| |
|
| | type=str, default=r"D:\Novalad\codes\eval\dp-bench\dataset\reference_table.json",
|
| |
|
| |
|
| | help="Path to the ground truth file"
|
| | )
|
| | parser.add_argument(
|
| | "--pred_path",
|
| | type=str, default=r"D:\Novalad\codes\eval\dp-bench\dataset\sample_results\novalad.json",
|
| | help="Path to the prediction file"
|
| | )
|
| | parser.add_argument(
|
| | "--ignore_classes_for_layout",
|
| | type=list, default=[],
|
| | help="List of layout classes to ignore. This is used only for layout evaluation."
|
| | )
|
| | parser.add_argument(
|
| | "--mode",
|
| | type=str, default="table",
|
| | help="Mode for evaluation (layout/table)"
|
| | )
|
| |
|
| | return parser.parse_args()
|
| |
|
| |
|
| | def main():
|
| | args = parse_args()
|
| |
|
| | print("Arguments:")
|
| | for k, v in vars(args).items():
|
| | print(f" {k}: {v}")
|
| | print("-" * 50)
|
| |
|
| | label_data = read_file(args.ref_path)
|
| | pred_data = read_file(args.pred_path)
|
| |
|
| | check_data_validity(label_data, pred_data)
|
| | if args.mode == "layout":
|
| | score = evaluate_layout(
|
| | label_data, pred_data,
|
| | ignore_classes=args.ignore_classes_for_layout,
|
| | )
|
| | print(f"NID Score: {score:.4f}")
|
| | elif args.mode == "table":
|
| | teds_score, teds_s_score = evaluate_table(label_data, pred_data)
|
| | print(f"TEDS Score: {teds_score:.4f}")
|
| | print(f"TEDS-S Score: {teds_s_score:.4f}")
|
| | else:
|
| | raise ValueError(f"{args.mode} mode not supported")
|
| |
|
| |
|
| | if __name__ == "__main__":
|
| | main()
|
| |
|