Spaces:
Runtime error
Runtime error
Add test_combined_models.py and compare/ folder (excluding cvat_project_7_export and Annika 2 folders)
0a216c0 | """ | |
| Test old models and new models on SampleBatch data. | |
| Compare both against ground truth and calculate metrics. | |
| Create side-by-side visualizations: GT | Old Models | New Models | |
| """ | |
| import os | |
| import sys | |
| import json | |
| import tempfile | |
| from pathlib import Path | |
| import numpy as np | |
| from PIL import Image | |
| import matplotlib.pyplot as plt | |
| import matplotlib.patches as patches | |
| # Add paths | |
| SCRIPT_DIR = os.path.dirname(os.path.abspath(__file__)) | |
| PROJECT_ROOT = os.path.dirname(os.path.dirname(SCRIPT_DIR)) | |
| sys.path.insert(0, SCRIPT_DIR) | |
| sys.path.insert(0, PROJECT_ROOT) | |
| from old_models import process_dataset as process_old_models, run_old_models_on_image | |
| from new_models import process_dataset as process_new_models, run_new_models_on_image | |
| from compare import calculate_metrics, align_categories, draw_coco_annotations_simple | |
| try: | |
| from pycocotools.coco import COCO | |
| from pycocotools.cocoeval import COCOeval | |
| HAS_PYCOCOTOOLS = True | |
| except ImportError: | |
| HAS_PYCOCOTOOLS = False | |
| print("Warning: pycocotools not available. Metrics calculation will be limited.") | |
| # No mapping - use old models' actual output directly | |
| # Sample batch folders | |
| SAMPLE_BATCH_FOLDERS = [ | |
| "SampleBatch2", | |
| "SampleBatch3", | |
| "SampleBatch4", | |
| ] | |
| # Removed mapping function - use old models' output directly | |
| def create_side_by_side_visualization(image_path, gt_coco, old_coco, new_coco, output_path): | |
| """ | |
| Create side-by-side visualization: GT | Old Models | New Models | |
| """ | |
| fig, axes = plt.subplots(1, 3, figsize=(30, 10)) | |
| # Left: Ground Truth | |
| draw_coco_annotations_simple(image_path, gt_coco, "Ground Truth", axes[0]) | |
| # Middle: Old Models | |
| draw_coco_annotations_simple(image_path, old_coco, "Old Models", axes[1]) | |
| # Right: New Models | |
| draw_coco_annotations_simple(image_path, new_coco, "New Models", axes[2]) | |
| plt.tight_layout() | |
| plt.savefig(output_path, dpi=150, bbox_inches='tight') | |
| plt.close() | |
| print(f" β Saved comparison to: {output_path}") | |
| def process_sample_batch(folder_name, base_dir=None): | |
| """ | |
| Process a single sample batch: test old and new models, calculate metrics, create visualizations. | |
| """ | |
| if base_dir is None: | |
| base_dir = SCRIPT_DIR | |
| folder_path = Path(base_dir) / folder_name | |
| if not folder_path.exists(): | |
| print(f"β οΈ Warning: Folder not found: {folder_path}") | |
| return None | |
| print("\n" + "=" * 70) | |
| print(f"Processing: {folder_name}") | |
| print("=" * 70) | |
| # Paths | |
| gt_json_path = folder_path / "Annotations" / "instances_default.json" | |
| images_dir = folder_path / "Images" | |
| output_dir = folder_path / "model_comparison" | |
| os.makedirs(output_dir, exist_ok=True) | |
| # Load ground truth | |
| print(f"\n[1/5] Loading ground truth...") | |
| with open(gt_json_path, 'r') as f: | |
| gt_coco = json.load(f) | |
| print(f" β Loaded {len(gt_coco['images'])} images") | |
| print(f" β Loaded {len(gt_coco['annotations'])} annotations") | |
| # Run old models | |
| print(f"\n[2/5] Running old models...") | |
| old_output_dir = output_dir / "old_models" | |
| os.makedirs(old_output_dir, exist_ok=True) | |
| try: | |
| old_coco = process_old_models(str(images_dir), str(old_output_dir)) | |
| print(f" β Generated {len(old_coco['annotations'])} annotations") | |
| print(f" β Categories: {[c['name'] for c in old_coco['categories']]}") | |
| except Exception as e: | |
| print(f" β Error running old models: {e}") | |
| import traceback | |
| traceback.print_exc() | |
| return None | |
| # Run new models | |
| print(f"\n[3/5] Running new models...") | |
| new_output_dir = output_dir / "new_models" | |
| os.makedirs(new_output_dir, exist_ok=True) | |
| try: | |
| new_coco = process_new_models(str(images_dir), str(new_output_dir)) | |
| print(f" β Generated {len(new_coco['annotations'])} annotations") | |
| except Exception as e: | |
| print(f" β Error running new models: {e}") | |
| import traceback | |
| traceback.print_exc() | |
| return None | |
| # Calculate metrics | |
| print(f"\n[4/5] Calculating metrics...") | |
| # Align categories with ground truth (by name matching) | |
| old_coco_aligned = align_categories(gt_coco.copy(), old_coco.copy()) | |
| new_coco_aligned = align_categories(gt_coco.copy(), new_coco.copy()) | |
| # Calculate metrics for old models | |
| print(f"\n Old Models Metrics:") | |
| old_metrics = calculate_metrics(gt_coco, old_coco_aligned, str(output_dir)) | |
| print(f" mAP@50: {old_metrics.get('mAP@50', 0):.4f}") | |
| print(f" mAP@[.50:.95]: {old_metrics.get('mAP@[.50:.95]', 0):.4f}") | |
| print(f" Precision: {old_metrics.get('Precision', 0):.4f}") | |
| print(f" Recall: {old_metrics.get('Recall', 0):.4f}") | |
| # Calculate metrics for new models | |
| print(f"\n New Models Metrics:") | |
| new_metrics = calculate_metrics(gt_coco, new_coco_aligned, str(output_dir)) | |
| print(f" mAP@50: {new_metrics.get('mAP@50', 0):.4f}") | |
| print(f" mAP@[.50:.95]: {new_metrics.get('mAP@[.50:.95]', 0):.4f}") | |
| print(f" Precision: {new_metrics.get('Precision', 0):.4f}") | |
| print(f" Recall: {new_metrics.get('Recall', 0):.4f}") | |
| # Save metrics | |
| metrics_path = output_dir / "metrics.json" | |
| with open(metrics_path, 'w') as f: | |
| json.dump({ | |
| "old_models": old_metrics, | |
| "new_models": new_metrics | |
| }, f, indent=4) | |
| print(f" β Saved metrics to: {metrics_path}") | |
| # Create visualizations | |
| print(f"\n[5/5] Creating side-by-side visualizations...") | |
| vis_dir = output_dir / "visualizations" | |
| os.makedirs(vis_dir, exist_ok=True) | |
| for img_info in gt_coco["images"]: | |
| image_name = img_info["file_name"] | |
| image_path = images_dir / image_name | |
| if not image_path.exists(): | |
| continue | |
| img_id = img_info["id"] | |
| # Filter annotations for this image | |
| gt_img_coco = { | |
| "images": [img_info], | |
| "annotations": [a for a in gt_coco["annotations"] if a["image_id"] == img_id], | |
| "categories": gt_coco["categories"] | |
| } | |
| old_img_coco = { | |
| "images": [img_info], | |
| "annotations": [a for a in old_coco["annotations"] if a["image_id"] == img_id], | |
| "categories": old_coco["categories"] | |
| } | |
| new_img_coco = { | |
| "images": [img_info], | |
| "annotations": [a for a in new_coco["annotations"] if a["image_id"] == img_id], | |
| "categories": new_coco["categories"] | |
| } | |
| # Create visualization | |
| output_path = vis_dir / f"{Path(image_name).stem}_comparison.png" | |
| create_side_by_side_visualization( | |
| str(image_path), | |
| gt_img_coco, | |
| old_img_coco, | |
| new_img_coco, | |
| str(output_path) | |
| ) | |
| print(f" β Visualizations saved to: {vis_dir}") | |
| return { | |
| "folder": folder_name, | |
| "old_metrics": old_metrics, | |
| "new_metrics": new_metrics, | |
| "old_annotations": len(old_coco["annotations"]), | |
| "new_annotations": len(new_coco["annotations"]), | |
| "gt_annotations": len(gt_coco["annotations"]) | |
| } | |
| def main(): | |
| """Main function to process all sample batches.""" | |
| print("=" * 70) | |
| print("TESTING MODELS ON SAMPLE BATCHES") | |
| print("=" * 70) | |
| print(f"\nProcessing {len(SAMPLE_BATCH_FOLDERS)} sample batch folders:") | |
| for folder in SAMPLE_BATCH_FOLDERS: | |
| print(f" - {folder}") | |
| results = [] | |
| for folder_name in SAMPLE_BATCH_FOLDERS: | |
| result = process_sample_batch(folder_name) | |
| if result: | |
| results.append(result) | |
| # Print summary | |
| print("\n" + "=" * 70) | |
| print("SUMMARY") | |
| print("=" * 70) | |
| for r in results: | |
| print(f"\n{r['folder']}:") | |
| print(f" Ground Truth: {r['gt_annotations']} annotations") | |
| print(f" Old Models: {r['old_annotations']} annotations") | |
| print(f" mAP@50: {r['old_metrics'].get('mAP@50', 0):.4f}") | |
| print(f" Precision: {r['old_metrics'].get('Precision', 0):.4f}") | |
| print(f" Recall: {r['old_metrics'].get('Recall', 0):.4f}") | |
| print(f" New Models: {r['new_annotations']} annotations") | |
| print(f" mAP@50: {r['new_metrics'].get('mAP@50', 0):.4f}") | |
| print(f" Precision: {r['new_metrics'].get('Precision', 0):.4f}") | |
| print(f" Recall: {r['new_metrics'].get('Recall', 0):.4f}") | |
| # Save summary | |
| summary_path = Path(SCRIPT_DIR) / "sample_batches_model_comparison_summary.json" | |
| with open(summary_path, 'w') as f: | |
| json.dump(results, f, indent=4) | |
| print(f"\nβ Summary saved to: {summary_path}") | |
| print("\n" + "=" * 70) | |
| print("COMPLETE!") | |
| print("=" * 70) | |
| if __name__ == "__main__": | |
| main() | |