monumenten / process_data.py
TomerGabay's picture
added process_data.py and requirements.txt
caa2ed1 verified
raw
history blame
7.66 kB
import asyncio
import glob
import json
import logging
import multiprocessing
import os
import xml.etree.ElementTree as ET
from datetime import datetime
from typing import List, Optional, Set
import pandas as pd
from datasets import Dataset
from dotenv import load_dotenv
from monumenten import MonumentenClient
load_dotenv()
# Configure logging
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
# Define constants
XML_DIRECTORY = "vbo_xmls/"
INTERMEDIATE_CSV_PATH = "verblijfsobjecten_ids.csv"
FINAL_CSV_PATH = "monumenten.csv"
HF_REPO_ID = "woonstadrotterdam/monumenten"
HF_TOKEN = os.getenv("HF_TOKEN")
def is_valid_identificatie(id_value: str) -> bool:
"""
Validate if the ID is a proper verblijfsobject ID.
Valid IDs must be 16 characters long, consist of digits,
and have '01' at positions 4-5 (0-indexed).
Example: 'xxxx01xxxxxxxxxx' where x is a digit (e.g., '0304010000269586').
"""
if id_value is None:
return False
return len(id_value) == 16 and id_value.isdigit() and id_value[4:6] == "01"
def extract_identificaties(xml_path: str) -> List[str]:
"""
Extract all valid identificatie values from a single XML file using iterative parsing.
"""
identificaties = []
try:
context = ET.iterparse(xml_path, events=("end",))
for event, elem in context:
if elem.tag.endswith("identificatie"):
id_value = elem.text
if is_valid_identificatie(id_value):
identificaties.append(id_value)
elem.clear() # Free memory
if identificaties:
logger.debug(
f"Found {len(identificaties)} valid identificatie values in {xml_path}"
)
return identificaties
except Exception as e:
logger.error(f"Error parsing XML file {xml_path}: {e}")
return []
def get_xml_files() -> List[str]:
"""
Get list of XML files from the specified directory.
"""
xml_files = glob.glob(os.path.join(XML_DIRECTORY, "*.xml"))
if not xml_files:
logger.error(f"No XML files found in {XML_DIRECTORY}")
else:
logger.info(f"Found {len(xml_files)} XML files in {XML_DIRECTORY}")
return xml_files
def process_files_parallel(xml_files: List[str]) -> Set[str]:
"""
Process XML files in parallel using multiprocessing.
Returns a set of unique identificaties.
"""
unique_identificaties = set()
logger.info(f"Starting parallel processing of {len(xml_files)} XML files...")
with multiprocessing.Pool() as pool:
results = pool.imap_unordered(extract_identificaties, xml_files)
for i, file_identificaties in enumerate(results):
unique_identificaties.update(file_identificaties)
if (i + 1) % 100 == 0: # Log progress every 100 files
logger.info(
f"Processed {i + 1}/{len(xml_files)} files. "
f"Current unique identificaties: {len(unique_identificaties)}"
)
logger.info(
f"All files processed. Total unique identificaties found: {len(unique_identificaties)}"
)
return unique_identificaties
def create_identificaties_dataframe(unique_ids: Set[str]) -> Optional[pd.DataFrame]:
"""
Create and save DataFrame from unique identificaties.
Returns the DataFrame or None if no valid identificaties found.
"""
if not unique_ids:
logger.info("No valid identificaties found.")
return None
df = pd.DataFrame(list(unique_ids), columns=["bag_verblijfsobject_id"])
logger.info(f"Created DataFrame with {len(df)} unique valid identificaties.")
# Save intermediate results
df.to_csv(INTERMEDIATE_CSV_PATH, index=False)
logger.info(f"Saved DataFrame to {INTERMEDIATE_CSV_PATH}")
# Display info
print("\nFirst few rows of the extracted identificaties DataFrame:")
print(df.head())
print("\nIdentificaties DataFrame Info:")
df.info()
return df
async def process_with_monumenten_client(df: pd.DataFrame) -> Optional[pd.DataFrame]:
"""
Process the DataFrame using MonumentenClient.
Returns processed DataFrame or None if processing fails.
"""
if df.empty:
logger.warning("Empty DataFrame provided to MonumentenClient.")
return None
logger.info(f"Processing {len(df)} identificaties with MonumentenClient...")
try:
async with MonumentenClient() as client:
result_df = await client.process_from_df(
df=df, verblijfsobject_id_col="bag_verblijfsobject_id"
)
logger.info("Finished processing with MonumentenClient.")
return result_df
except Exception as e:
logger.error(f"Error processing with MonumentenClient: {e}")
return None
def save_final_results(result_df: Optional[pd.DataFrame]) -> None:
"""
Save the final results to CSV if valid data is present.
"""
if result_df is not None and not result_df.empty:
result_df.to_csv(FINAL_CSV_PATH, index=False)
logger.info(f"Successfully saved final monumenten data to {FINAL_CSV_PATH}")
print(f"\nFinal data saved to {FINAL_CSV_PATH}")
print(result_df.head())
# Push to Hugging Face
if push_to_huggingface(result_df):
print(f"\nData successfully pushed to Hugging Face dataset: {HF_REPO_ID}")
else:
print("\nFailed to push data to Hugging Face. Check logs for details.")
elif result_df is not None and result_df.empty:
logger.info("Processing resulted in an empty DataFrame. Nothing to save.")
print("\nProcessing resulted in an empty DataFrame.")
else:
logger.warning("No valid data to save. Process did not complete successfully.")
print("\nProcess did not complete successfully or returned no data.")
def push_to_huggingface(result_df: pd.DataFrame) -> bool:
"""
Push the final results to Hugging Face datasets hub using datasets.push_to_hub
with a custom split name.
Returns True if successful, False otherwise.
"""
if not HF_TOKEN:
logger.error("No Hugging Face token found in environment variables (HF_TOKEN)")
return False
if result_df.empty:
logger.warning(
"Result DataFrame is empty. Skipping push of main dataset to Hugging Face."
)
else:
logger.info(
f"Converting DataFrame with {len(result_df)} rows to Hugging Face Dataset."
)
hf_dataset_single = Dataset.from_pandas(result_df)
hf_dataset_single.push_to_hub(
repo_id=HF_REPO_ID,
commit_message=f"Update monumenten dataset",
token=HF_TOKEN,
)
logger.info(f"Successfully pushed dataset dictionary to {HF_REPO_ID}")
async def main() -> Optional[pd.DataFrame]:
"""
Main function orchestrating the entire process.
Returns the final processed DataFrame or None if processing fails.
"""
# Get XML files
xml_files = get_xml_files()
if not xml_files:
return None
# Process files and get unique identificaties
unique_identificaties = process_files_parallel(xml_files)
# Create DataFrame from unique identificaties
df = create_identificaties_dataframe(unique_identificaties)
if df is None:
return None
# Process with MonumentenClient
result_df = await process_with_monumenten_client(df)
return result_df
if __name__ == "__main__":
# Run main process
result_dataframe = asyncio.run(main())
# Save results
save_final_results(result_dataframe)