Datasets:

Modalities:
Image
Languages:
English
Size:
< 1K
Libraries:
Datasets
License:
IOAI-ISC's picture
Tasks Restroom Added
abcc75c verified
import numpy as np
import json
import os
import sys
def calculate_precision_at_1(predictions, ground_truth):
"""
Calculate precision@1 for the given predictions and ground truth.
Args:
predictions (np.array): Array of predicted gallery indices for each query
ground_truth (np.array): Array of correct gallery indices for each query
Returns:
float: Precision@1 score (percentage of correct predictions)
"""
if len(predictions) != len(ground_truth):
raise ValueError(f"Predictions length ({len(predictions)}) doesn't match ground truth length ({len(ground_truth)})")
# Count correct predictions
correct_predictions = np.sum(predictions == ground_truth)
total_predictions = len(predictions)
# Calculate precision@1 as percentage
precision_at_1 = (correct_predictions / total_predictions)
return precision_at_1
def load_submission_file(filepath):
"""
Load submission file and handle potential errors.
Args:
filepath (str): Path to the submission file
Returns:
np.array or None: Loaded array or None if file doesn't exist or is invalid
"""
try:
if not os.path.exists(filepath):
print(f"Warning: Submission file {filepath} not found")
return None
submission = np.load(filepath)
print(f"Loaded {filepath}: shape {submission.shape}, dtype {submission.dtype}")
return submission
except Exception as e:
print(f"Error loading {filepath}: {str(e)}")
return None
def load_ground_truth_file(filepath):
"""
Load ground truth file.
Args:
filepath (str): Path to the ground truth file
Returns:
np.array: Loaded ground truth array
"""
try:
if not os.path.exists(filepath):
raise FileNotFoundError(f"Ground truth file {filepath} not found")
ground_truth = np.load(filepath)
print(f"Loaded ground truth {filepath}: shape {ground_truth.shape}, dtype {ground_truth.dtype}")
return ground_truth
except Exception as e:
print(f"Error loading ground truth {filepath}: {str(e)}")
raise
def evaluate_test_set(submission_file, ground_truth_file, test_name):
"""
Evaluate a single test set.
Args:
submission_file (str): Path to submission file
ground_truth_file (str): Path to ground truth file
test_name (str): Name of the test set for logging
Returns:
float or None: Precision@1 score or None if evaluation failed
"""
print(f"\n=== Evaluating {test_name} ===")
# Load ground truth
try:
ground_truth = load_ground_truth_file(ground_truth_file)
except Exception as e:
print(f"Failed to load ground truth for {test_name}: {str(e)}")
return None
# Load submission
submission = load_submission_file(submission_file)
if submission is None:
print(f"Failed to load submission for {test_name}")
return None
# Validate submission format
if submission.shape != ground_truth.shape:
print(f"Shape mismatch for {test_name}: submission {submission.shape} vs ground truth {ground_truth.shape}")
return None
# Calculate precision@1
try:
score = calculate_precision_at_1(submission, ground_truth)
print(f"{test_name} - Precision@1: {score:.2f}")
# Log some statistics
correct_count = np.sum(submission == ground_truth)
total_count = len(submission)
print(f"{test_name} - Correct predictions: {correct_count}/{total_count}")
return score
except Exception as e:
print(f"Error calculating precision@1 for {test_name}: {str(e)}")
return None
def main():
"""
Main evaluation function.
"""
print("Starting evaluation...")
if os.environ.get('METRIC_PATH'):
METRIC_PATH = os.environ.get("METRIC_PATH") + "/"
else:
METRIC_PATH = "" # Fallback for local testing
# File paths
submission_a_file = "submission_a.npy"
submission_b_file = "submission_b.npy"
ground_truth_a_file = METRIC_PATH + "answer_a.npy"
ground_truth_b_file = METRIC_PATH + "answer_b.npy"
output_file = "score.json"
# Evaluate test set A
score_a = evaluate_test_set(submission_a_file, ground_truth_a_file, "Test Set A")
# Evaluate test set B
score_b = evaluate_test_set(submission_b_file, ground_truth_b_file, "Test Set B")
# Determine overall status
status = True
msg = "Success!"
# Handle missing or failed evaluations
if score_a is None:
score_a = 0.0
status = False
msg = "Failed to evaluate Test Set A"
if score_b is None:
score_b = 0.0
if status: # Only update if not already failed
status = False
msg = "Failed to evaluate Test Set B"
else:
msg = "Failed to evaluate both test sets"
if score_a > 1:
score_a = 0.0
if score_b > 1:
score_b = 0.0
def sanitize_score(value):
"""处理单个分数值,将NaN和inf替换为0"""
if not np.isfinite(value):
return 0.0
return value
# Create result dictionary
result = {
"status": status,
"score": {
"public_a": sanitize_score(score_a),
"private_b": sanitize_score(score_b),
},
"msg": msg,
}
# Save results to JSON
try:
with open(output_file, 'w') as f:
json.dump(result, f, indent=4)
print(f"\nResults saved to {output_file}")
except Exception as e:
print(f"Error saving results to {output_file}: {str(e)}")
sys.exit(1)
# Print final summary
print("\n=== EVALUATION SUMMARY ===")
print(f"Status: {status}")
print(f"Test Set A (public) Score: {score_a:.2f}")
print(f"Test Set B (private) Score: {score_b:.2f}")
print(f"Message: {msg}")
return result
if __name__ == "__main__":
main()