| import argparse |
| import sys |
| sys.path.append(r"C:\Users\FCI\Desktop\engineero_ai\pg-clean-search") |
| from app.cleaning.pipeline import run_cleaning_from_yaml |
|
|
| def main(): |
| """Defines the command-line interface for the cleaning pipeline.""" |
| p = argparse.ArgumentParser( |
| description="Run an in-place data cleaning pipeline on specified PostgreSQL tables.", |
| formatter_class=argparse.RawTextHelpFormatter |
| ) |
| p.add_argument( |
| "--cfg", |
| |
| default="scripts\clean_test.yaml", |
| help="YAML file listing source schema, table names, primary keys, and culprit columns." |
| ) |
| p.add_argument( |
| "--batch", |
| type=int, |
| default=1000, |
| help=( |
| "If --clean-all=False, this is the maximum number of rows read (LIMIT) and processed (1 batch).\n" |
| "If --clean-all=True, this is the chunk size for reading/writing multiple batches." |
| ) |
| ) |
| p.add_argument( |
| "--clean-all", |
| action="store_true", |
| help="Process and clean every row in the table (overrides the LIMIT set by --batch)." |
| ) |
| p.add_argument( |
| "--clean_cap", |
| type=int, |
| |
| help="Maximum number of rows to clean per table." |
| ) |
| |
| args = p.parse_args() |
| |
| run_cleaning_from_yaml( |
| args.cfg, |
| batch_size=args.batch, |
| clean_all=args.clean_all, |
| clean_cap=args.clean_cap, |
| ) |
|
|
| if __name__ == "__main__": |
| main() |
| |