danielrosehill's picture
Add parametric analysis and dataset improvements with LFS tracking
04edfe7
#!/usr/bin/env python3
"""
Parametric analysis of geodata mapping in IFVI Global Value Factors Dataset
Analyzes the geographic entity distribution and ISO code coverage
"""
import json
import os
from collections import Counter
def analyze_geodata():
"""
Analyze the geodata.json file for geographic statistics
"""
geodata_path = "/home/daniel/repos/hugging-face/IFVI-Global-Value-Factors-Dataset-V2/processing/remapping/geodata.json"
output_dir = "/home/daniel/repos/hugging-face/IFVI-Global-Value-Factors-Dataset-V2/parametric-data"
print("IFVI Global Value Factors Dataset - Geographic Entity Analysis")
print("=" * 65)
print()
try:
with open(geodata_path, 'r', encoding='utf-8') as f:
data = json.load(f)
metadata = data.get('metadata', {})
mapping = data.get('mapping', [])
# Extract statistics from metadata
total_entities = metadata.get('total_entities', 0)
entity_types = metadata.get('entity_types', {})
entities_with_iso = metadata.get('entities_with_iso', 0)
us_states_with_codes = metadata.get('us_states_with_codes', 0)
# Calculate additional statistics
entities_without_iso = total_entities - entities_with_iso - us_states_with_codes
# Analyze regions
region_counts = Counter()
iso_by_region = Counter()
non_iso_by_region = Counter()
for entity in mapping:
region = entity.get('region', 'Unknown')
has_iso = entity.get('has_iso', False)
entity_type = entity.get('entity_type', 'unknown')
region_counts[region] += 1
if has_iso:
iso_by_region[region] += 1
elif entity_type != 'us_state': # Don't count US states as "without ISO"
non_iso_by_region[region] += 1
# Display results
print("GEOGRAPHIC ENTITY STATISTICS")
print("-" * 40)
print(f"Total unique geolocations: {total_entities:,}")
print(f"Entities with ISO 3166-1 codes: {entities_with_iso:,}")
print(f"US states (with state codes): {us_states_with_codes:,}")
print(f"Non-sovereign entities (no ISO): {entities_without_iso:,}")
print()
print("ENTITY TYPE BREAKDOWN")
print("-" * 25)
for entity_type, count in entity_types.items():
percentage = (count / total_entities) * 100
print(f"{entity_type.replace('_', ' ').title()}: {count:,} ({percentage:.1f}%)")
print()
print("ISO CODE COVERAGE")
print("-" * 20)
total_with_codes = entities_with_iso + us_states_with_codes
coverage_percentage = (total_with_codes / total_entities) * 100
print(f"Entities with standardized codes: {total_with_codes:,} ({coverage_percentage:.1f}%)")
print(f"Entities without codes: {entities_without_iso:,} ({(entities_without_iso/total_entities)*100:.1f}%)")
print()
print("REGIONAL DISTRIBUTION")
print("-" * 22)
for region in sorted(region_counts.keys()):
total_in_region = region_counts[region]
with_iso = iso_by_region.get(region, 0)
without_iso = non_iso_by_region.get(region, 0)
print(f"{region}:")
print(f" Total entities: {total_in_region}")
print(f" With ISO codes: {with_iso}")
print(f" Without ISO codes: {without_iso}")
print()
# Create structured output
analysis_results = {
"analysis_metadata": {
"timestamp": "2025-08-21T21:20:00+03:00",
"source_file": "processing/remapping/geodata.json",
"iso_standard": metadata.get('iso_standard', 'ISO 3166-1 alpha-3')
},
"geographic_statistics": {
"total_unique_geolocations": total_entities,
"entities_with_iso_codes": entities_with_iso,
"us_states_with_codes": us_states_with_codes,
"non_sovereign_entities": entities_without_iso,
"total_with_standardized_codes": total_with_codes,
"code_coverage_percentage": round(coverage_percentage, 2)
},
"entity_type_breakdown": {
entity_type: {
"count": count,
"percentage": round((count / total_entities) * 100, 2)
}
for entity_type, count in entity_types.items()
},
"regional_distribution": {
region: {
"total_entities": region_counts[region],
"with_iso_codes": iso_by_region.get(region, 0),
"without_iso_codes": non_iso_by_region.get(region, 0)
}
for region in sorted(region_counts.keys())
}
}
# Save to JSON
output_file = os.path.join(output_dir, "geodata_analysis.json")
with open(output_file, 'w', encoding='utf-8') as f:
json.dump(analysis_results, f, indent=2, ensure_ascii=False)
print(f"Results saved to: {output_file}")
# Save summary CSV
import csv
csv_file = os.path.join(output_dir, "geodata_summary.csv")
with open(csv_file, 'w', newline='', encoding='utf-8') as f:
writer = csv.writer(f)
writer.writerow(["Metric", "Count", "Percentage"])
writer.writerow(["Total Geolocations", total_entities, "100.0"])
writer.writerow(["With ISO Codes", entities_with_iso, f"{(entities_with_iso/total_entities)*100:.1f}"])
writer.writerow(["US States", us_states_with_codes, f"{(us_states_with_codes/total_entities)*100:.1f}"])
writer.writerow(["Non-Sovereign", entities_without_iso, f"{(entities_without_iso/total_entities)*100:.1f}"])
print(f"Summary CSV saved to: {csv_file}")
except Exception as e:
print(f"Error analyzing geodata: {e}")
if __name__ == "__main__":
analyze_geodata()