File size: 5,709 Bytes
2362e57 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 | #!/usr/bin/env python3
"""
Cleanup script to remove test/dummy submissions from leaderboard.
Usage:
python cleanup_test_data.py --model-name "TestModel"
python cleanup_test_data.py --all-test # Remove all models with "test" in name
python cleanup_test_data.py --list # List all models
"""
import json
import argparse
import shutil
from pathlib import Path
# Configuration
PERSISTENT_DIR = Path("/data") if Path("/data").exists() else Path(".")
LEADERBOARD_FILE = PERSISTENT_DIR / "leaderboard.json"
RESULTS_DIR = PERSISTENT_DIR / "results"
SUBMISSIONS_DIR = PERSISTENT_DIR / "submissions"
def load_leaderboard():
"""Load leaderboard from file."""
if not LEADERBOARD_FILE.exists():
print(f"❌ Leaderboard file not found: {LEADERBOARD_FILE}")
return []
with open(LEADERBOARD_FILE, 'r') as f:
data = json.load(f)
return data if isinstance(data, list) else []
def save_leaderboard(data):
"""Save leaderboard to file."""
# Update ranks
for i, entry in enumerate(data, 1):
entry['rank'] = i
with open(LEADERBOARD_FILE, 'w') as f:
json.dump(data, f, indent=2)
print(f"✓ Saved leaderboard with {len(data)} entries")
def list_models():
"""List all models in leaderboard."""
data = load_leaderboard()
if not data:
print("Leaderboard is empty")
return
print(f"\n{'='*80}")
print(f"LEADERBOARD MODELS ({len(data)} total)")
print(f"{'='*80}\n")
for entry in data:
rank = entry.get('rank', '?')
model_name = entry.get('model_name', 'Unknown')
organization = entry.get('organization', 'Unknown')
date = entry.get('date', 'Unknown')
print(f"#{rank}: {model_name}")
print(f" Organization: {organization}")
print(f" Date: {date}")
print()
def delete_model(model_name, dry_run=False):
"""Delete a model from leaderboard and cleanup associated files."""
data = load_leaderboard()
# Find the model
model_entry = None
for entry in data:
if entry.get('model_name') == model_name:
model_entry = entry
break
if not model_entry:
print(f"❌ Model not found: {model_name}")
return False
print(f"\n{'='*80}")
print(f"DELETING MODEL: {model_name}")
print(f"{'='*80}\n")
print(f"Organization: {model_entry.get('organization')}")
print(f"Date: {model_entry.get('date')}")
print(f"Rank: {model_entry.get('rank')}")
if dry_run:
print("\n⚠️ DRY RUN - No changes will be made")
# Associated files/directories
model_dir_name = model_name.replace(" ", "_")
results_dir = RESULTS_DIR / model_dir_name
# Remove from leaderboard
if not dry_run:
data = [e for e in data if e.get('model_name') != model_name]
save_leaderboard(data)
print(f"✓ Removed from leaderboard")
else:
print(f"[DRY RUN] Would remove from leaderboard")
# Remove results directory
if results_dir.exists():
if not dry_run:
shutil.rmtree(results_dir)
print(f"✓ Removed results directory: {results_dir}")
else:
print(f"[DRY RUN] Would remove: {results_dir}")
# Check for submission files (might not exist after evaluation)
submission_dir = SUBMISSIONS_DIR / model_dir_name
if submission_dir.exists():
if not dry_run:
shutil.rmtree(submission_dir)
print(f"✓ Removed submissions directory: {submission_dir}")
else:
print(f"[DRY RUN] Would remove: {submission_dir}")
print(f"\n{'='*80}")
if dry_run:
print("✓ DRY RUN COMPLETE - No changes made")
else:
print("✓ MODEL DELETED SUCCESSFULLY")
print(f"{'='*80}\n")
return True
def delete_test_models(dry_run=False):
"""Delete all models with 'test' in name (case-insensitive)."""
data = load_leaderboard()
test_models = [
entry for entry in data
if 'test' in entry.get('model_name', '').lower() or
'test' in entry.get('organization', '').lower()
]
if not test_models:
print("No test models found")
return
print(f"\n{'='*80}")
print(f"FOUND {len(test_models)} TEST MODELS")
print(f"{'='*80}\n")
for entry in test_models:
print(f"- {entry.get('model_name')} ({entry.get('organization')})")
print()
if not dry_run:
confirm = input("Delete all these models? (yes/no): ")
if confirm.lower() != 'yes':
print("Cancelled")
return
for entry in test_models:
model_name = entry.get('model_name')
delete_model(model_name, dry_run=dry_run)
def main():
parser = argparse.ArgumentParser(
description="Cleanup test/dummy submissions from MedVidBench leaderboard"
)
group = parser.add_mutually_exclusive_group(required=True)
group.add_argument("--list", action="store_true",
help="List all models in leaderboard")
group.add_argument("--model-name", type=str,
help="Delete specific model by name")
group.add_argument("--all-test", action="store_true",
help="Delete all models with 'test' in name")
parser.add_argument("--dry-run", action="store_true",
help="Show what would be deleted without actually deleting")
args = parser.parse_args()
if args.list:
list_models()
elif args.model_name:
delete_model(args.model_name, dry_run=args.dry_run)
elif args.all_test:
delete_test_models(dry_run=args.dry_run)
if __name__ == "__main__":
main()
|