#!/usr/bin/env python3 """ fetch_crs.py — Download CRs and TSs from a 3GPP/ETSI Excel contribution list. Usage: python3 fetch_crs.py [--output-dir DIR] Steps: 1. Parse Excel, filter Accepted CRs by person name 2. Download CR DOCXs via ETSI docbox 3. Parse CR cover pages to extract target TS spec + version 4. Download TS DOCXs via ETSI portal WKI chain 5. Print summary report """ import argparse import os import re import sys import zipfile from pathlib import Path from etsi_client import CRFetcher, TSFetcher # --------------------------------------------------------------------------- # Path helpers # --------------------------------------------------------------------------- def wsl_path(p: str) -> str: """Convert Windows path (C:\\...) to WSL path (/mnt/c/...) if needed.""" p = p.strip() if len(p) >= 2 and p[1] == ":" and p[0].isalpha(): drive = p[0].lower() rest = p[2:].replace("\\", "/") return f"/mnt/{drive}{rest}" return p # --------------------------------------------------------------------------- # Step 1 — Parse Excel # --------------------------------------------------------------------------- def parse_excel(excel_path: str, person_name: str): """ Return list of (uid, title) for Accepted CRs matching person_name. Handles both .xls and .xlsx. """ path = Path(wsl_path(excel_path)) ext = path.suffix.lower() if ext == ".xls": return _parse_xls(path, person_name) elif ext == ".xlsx": return _parse_xlsx(path, person_name) else: raise ValueError(f"Unsupported file extension: {ext!r}. Expected .xls or .xlsx") def parse_excel_all_accepted(excel_path: str): """ Return list of (uid, title, submitted_by) for ALL Accepted CRs regardless of who submitted them. Used by build_cr_index.py. Handles both .xls and .xlsx. """ path = Path(wsl_path(excel_path)) ext = path.suffix.lower() if ext == ".xls": return _parse_xls(path) elif ext == ".xlsx": return _parse_xlsx(path) else: raise ValueError(f"Unsupported file extension: {ext!r}. Expected .xls or .xlsx") def _name_pattern(name: str) -> re.Pattern: return re.compile(r"\b" + re.escape(name) + r"\b", re.IGNORECASE) def _parse_xls(path: Path, person_name: str | None = None): """ Return Accepted CRs from an .xls file. If person_name is given: return (uid, title) filtered to that name. If person_name is None: return (uid, title, submitted_by) for all. """ try: import xlrd except ImportError: sys.exit("ERROR: xlrd is not installed. Run: pip install xlrd") wb = xlrd.open_workbook(str(path)) # Try "Contributions" sheet first, fall back to first sheet try: ws = wb.sheet_by_name("Contributions") except xlrd.XLRDError: ws = wb.sheet_by_index(0) # Row 0 is headers; row 1 is an empty duplicate — skip it headers = [str(ws.cell_value(0, c)).strip() for c in range(ws.ncols)] col = {h: i for i, h in enumerate(headers)} uid_col = col.get("Uid") or col.get("UID") or col.get("uid") type_col = col.get("Type") or col.get("type") status_col = col.get("Status") or col.get("status") by_col = col.get("SubmittedBy") or col.get("Submitted By") or col.get("submittedby") title_col = col.get("Title") or col.get("title") for name, c, required in [ ("Uid", uid_col, True), ("Type", type_col, True), ("Status", status_col, True), ("SubmittedBy", by_col, True), ("Title", title_col, False), ]: if c is None and required: raise ValueError(f"Column {name!r} not found. Available: {list(col.keys())}") pattern = _name_pattern(person_name) if person_name else None results = [] for r in range(2, ws.nrows): # skip header + empty duplicate uid = str(ws.cell_value(r, uid_col)).strip() doc_type = str(ws.cell_value(r, type_col)).strip() status = str(ws.cell_value(r, status_col)).strip() submitted_by = str(ws.cell_value(r, by_col)).strip() title = str(ws.cell_value(r, title_col)).strip() if title_col is not None else "" if doc_type != "CR": continue if status != "Accepted": continue if pattern and not pattern.search(submitted_by): continue results.append((uid, title) if person_name is not None else (uid, title, submitted_by)) return results def _parse_xlsx(path: Path, person_name: str | None = None): """ Return Accepted CRs from an .xlsx file. If person_name is given: return (uid, title) filtered to that name. If person_name is None: return (uid, title, submitted_by) for all. """ try: import openpyxl except ImportError: sys.exit("ERROR: openpyxl is not installed. Run: pip install openpyxl") wb = openpyxl.load_workbook(str(path), read_only=True, data_only=True) ws = wb["Contributions"] if "Contributions" in wb.sheetnames else wb.active rows = iter(ws.iter_rows(values_only=True)) # Row 0: headers header_row = next(rows) headers = [str(h).strip() if h is not None else "" for h in header_row] col = {h: i for i, h in enumerate(headers)} # Row 1: empty duplicate — skip next(rows, None) uid_col = col.get("Uid") or col.get("UID") or col.get("uid") type_col = col.get("Type") or col.get("type") status_col = col.get("Status") or col.get("status") by_col = col.get("SubmittedBy") or col.get("Submitted By") or col.get("submittedby") title_col = col.get("Title") or col.get("title") for name, c, required in [ ("Uid", uid_col, True), ("Type", type_col, True), ("Status", status_col, True), ("SubmittedBy", by_col, True), ("Title", title_col, False), ]: if c is None and required: raise ValueError(f"Column {name!r} not found. Available: {list(col.keys())}") pattern = _name_pattern(person_name) if person_name else None results = [] for row in rows: def cell(c): v = row[c] if c < len(row) else None return str(v).strip() if v is not None else "" uid = cell(uid_col) doc_type = cell(type_col) status = cell(status_col) submitted_by = cell(by_col) title = cell(title_col) if title_col is not None else "" if not uid: continue if doc_type != "CR": continue if status != "Accepted": continue if pattern and not pattern.search(submitted_by): continue results.append((uid, title) if person_name is not None else (uid, title, submitted_by)) return results # --------------------------------------------------------------------------- # Step 2 — Download CR DOCXs # --------------------------------------------------------------------------- def download_cr(uid: str, cr_dir: Path, eol_user: str, eol_password: str): """ Download CR DOCX for the given UID. Returns: (docx_path, note) — docx_path is the file to use for parsing note is a human-readable string for the summary Returns (None, error_msg) on failure. """ dest = cr_dir / f"{uid}.docx" if dest.exists(): extracted = cr_dir / f"{uid}_extracted.docx" if extracted.exists(): return extracted, "already existed" return dest, "already existed" try: finder = CRFetcher(eol_user, eol_password) url = finder.search_document(uid) if isinstance(url, str) and "not found" in url.lower(): return None, f"document not found: {uid}" content = finder.download_document(url) except Exception as e: return None, f"download error: {e}" if not content: return None, "empty response" dest.write_bytes(content) # ZIP detection if content[:4] == b"PK\x03\x04": try: with zipfile.ZipFile(dest) as zf: docx_entries = [n for n in zf.namelist() if n.endswith(".docx")] if docx_entries: extracted_name = f"{uid}_extracted.docx" extracted_path = cr_dir / extracted_name with zf.open(docx_entries[0]) as src, open(extracted_path, "wb") as dst: dst.write(src.read()) return extracted_path, "extracted from ZIP" except zipfile.BadZipFile: pass # Not actually a ZIP despite magic bytes — treat as raw DOCX return dest, "downloaded" # --------------------------------------------------------------------------- # Step 3 — Parse CR Cover Pages # --------------------------------------------------------------------------- SPEC_PATTERN = re.compile(r"^\d{3}\s\d{3}(-\d+)*$") # "102 221" or "102 230-2" SPEC_SEARCH = re.compile(r"\b\d{3}\s\d{3}(?:-\d+)*\b") # substring search fallback VERSION_PATTERN = re.compile(r"^[Vv]?\d+\.\d+(\.\d+)?$") # X.Y or X.Y.Z, optional V prefix VERSION_SEARCH = re.compile(r"\b\d+\.\d+\.\d+\b") # substring fallback def _normalise_version(v: str) -> str: """Strip optional V prefix and pad to X.Y.Z.""" v = v.lstrip('Vv') parts = v.split('.') while len(parts) < 3: parts.append('0') return '.'.join(parts[:3]) def _find_cover_table(doc): """Return the CR cover table, scanning all tables for one containing CHANGE REQUEST.""" MARKERS = {"CHANGE REQUEST", "CR", "CHANGE REQUEST"} for tbl in doc.tables: cells_text = {c.text.strip() for row in tbl.rows for c in row.cells} if cells_text & MARKERS: return tbl return None def parse_cr_cover(docx_path: Path): """ Parse the CR cover table (tables[0]) to extract (spec_number, version). Returns (spec_number, version) e.g. ("102 221", "18.3.0") Returns (None, None) if parsing fails. """ try: from docx import Document except ImportError: sys.exit("ERROR: python-docx is not installed. Run: pip install python-docx") try: doc = Document(str(docx_path)) except Exception as e: return None, None if not doc.tables: return None, None table = _find_cover_table(doc) if table is None: return None, None # Collect all non-empty cell texts in order cells = [] for row in table.rows: for cell in row.cells: text = cell.text.strip() if text: cells.append(text) cells = list(dict.fromkeys(cells)) spec_number = None version = None for i, text in enumerate(cells): # ── Strategy 1: exact cell match "NNN NNN" or "NNN NNN-N" ──────────── if spec_number is None and SPEC_PATTERN.match(text): spec_number = text # ── Strategy 2: positional — cell immediately after "CHANGE REQUEST" ─ # The cover table always places the spec number in the cell right after # the "CHANGE REQUEST" label. if spec_number is None and text.strip() == "CHANGE REQUEST" and i + 1 < len(cells): candidate = cells[i + 1].strip() if SPEC_PATTERN.match(candidate): spec_number = candidate # ── Version: cell immediately after "Current version:" ─────────────── if "Current version:" in text and i + 1 < len(cells): candidate = cells[i + 1] if VERSION_PATTERN.match(candidate): version = _normalise_version(candidate) if text in ("Current version:", "Current version") and version is None: if i + 1 < len(cells) and VERSION_PATTERN.match(cells[i + 1]): version = _normalise_version(cells[i + 1]) # ── Strategy 3: substring search across all cells ───────────────────────── # Catches cases where the spec number is embedded in a longer cell string. if spec_number is None: for text in cells: m = SPEC_SEARCH.search(text) if m: spec_number = m.group(0) break if spec_number: spec_number = spec_number.replace('\xa0', ' ').strip() return spec_number, version # --------------------------------------------------------------------------- # Step 4 — Download TS DOCXs # --------------------------------------------------------------------------- def download_ts(spec_number: str, version: str, ts_dir: Path, eol_user: str = "", eol_password: str = ""): """ Download TS DOCX for spec_number (e.g. "102 221") and version (e.g. "18.3.0"). Returns (filename, note) or (None, error_msg). """ spec_no_space = spec_number.replace(" ", "") filename = f"ts_{spec_no_space}_v{version}.docx" dest = ts_dir / filename if dest.exists(): return filename, "already existed" try: finder = TSFetcher(eol_user, eol_password) tmp_path = finder.search_document_docx(spec_number, version) except Exception as e: return None, f"download error: {e}" if "not found" in str(tmp_path).lower() or "rejected" in str(tmp_path).lower(): return None, tmp_path content = Path(tmp_path).read_bytes() if not content: return None, "empty response" dest.write_bytes(content) if content[:2] != b"PK": dest.unlink() return None, f"invalid file (not a ZIP/DOCX, starts with {content[:4]!r})" # Verify the TS contains the expected spec number AND version in its first paragraph try: import docx as _docx _doc = _docx.Document(dest) first_para = _doc.paragraphs[0].text if _doc.paragraphs else "" if spec_no_space not in first_para.replace(" ", ""): dest.unlink() return None, ( f"wrong TS returned: got {first_para[:80]!r} " f"(expected spec {spec_no_space})" ) if f"V{version}" not in first_para: dest.unlink() return None, ( f"wrong version returned: got {first_para[:80]!r} " f"(expected V{version})" ) except Exception: pass # Trust the ZIP check above return filename, "downloaded" # --------------------------------------------------------------------------- # Main # --------------------------------------------------------------------------- def main(): parser = argparse.ArgumentParser( description="Download CRs and TSs from a 3GPP/ETSI Excel contribution list." ) parser.add_argument("excel_path", help="Path to .xls or .xlsx contribution list") parser.add_argument("person_name", help="Name to search for in SubmittedBy column") parser.add_argument( "--output-dir", default=str(Path.home() / "CR_Processing"), help="Base output directory (default: ~/CR_Processing)", ) args = parser.parse_args() excel_path = wsl_path(args.excel_path) person_name = args.person_name output_dir = Path(wsl_path(args.output_dir)).expanduser() eol_user = os.environ.get("EOL_USER", "") eol_password = os.environ.get("EOL_PASSWORD", "") if not eol_user or not eol_password: sys.exit("ERROR: EOL_USER and EOL_PASSWORD must be set") cr_dir = output_dir / "CRs" ts_dir = output_dir / "TS" cr_dir.mkdir(parents=True, exist_ok=True) ts_dir.mkdir(parents=True, exist_ok=True) # --- Step 1: Parse Excel --- print(f"Parsing Excel: {excel_path}") print(f"Filtering for: {person_name!r} | Type=CR | Status=Accepted\n") try: cr_list = parse_excel(excel_path, person_name) except Exception as e: sys.exit(f"ERROR parsing Excel: {e}") print(f"Found {len(cr_list)} matching CR(s).\n") if not cr_list: print("Nothing to download.") return # --- Step 2: Download CR DOCXs --- print("Downloading CRs...") cr_results = [] # list of (uid, docx_path_or_None, note) for uid, title in cr_list: #print(f" [{uid}] ", end="", flush=True) docx_path, note = download_cr(uid, cr_dir, eol_user, eol_password) cr_results.append((uid, docx_path, note)) if docx_path: print(f"OK ({note}) — {docx_path.name}") else: print(f"FAILED — {note}") print() # --- Step 3: Parse cover pages --- print("Parsing CR cover pages...") ts_targets = {} # (spec_number, version) -> list of uids for uid, docx_path, note in cr_results: if docx_path is None: continue spec_number, version = parse_cr_cover(docx_path) if spec_number and version: key = (spec_number, version) ts_targets.setdefault(key, []).append(uid) print(f" [{uid}] → TS {spec_number} v{version}") else: print(f" [{uid}] WARNING: could not parse cover page (spec/version not found)") print() # --- Step 4: Download TSs --- print("Downloading TSs...") ts_results = [] # list of (spec_number, version, filename_or_None, note) for (spec_number, version), uids in ts_targets.items(): print(f" [TS {spec_number} v{version}] ", end="", flush=True) filename, note = download_ts(spec_number, version, ts_dir, eol_user, eol_password) ts_results.append((spec_number, version, filename, note)) if filename: print(f"OK ({note}) — {filename}") else: print(f"FAILED — {note}") print() # --- Step 5: Summary --- print("=" * 50) print("=== fetch-crs summary ===") print(f"Person: {person_name}") print(f"Excel: {excel_path}") print(f"CRs found: {len(cr_list)} (Accepted, Type=CR)") print() print("CRs downloaded:") for uid, docx_path, note in cr_results: if docx_path: print(f" ✓ {docx_path.name} [{note}]") else: print(f" ✗ {uid} — {note}") print() print("TSs downloaded:") for spec_number, version, filename, note in ts_results: if filename: print(f" ✓ {filename} [{note}]") else: print(f" ✗ ts_{spec_number.replace(' ', '')} v{version} — {note}") print() print(f"Output: {output_dir}/") if __name__ == "__main__": main()