| | |
| | """ |
| | Validate Interslavic MT corpus format. |
| | - Monolingual: id, sentence, source; no duplicate ids. |
| | - Parallel: id, isv, target, translator, method, review_status; ids exist in monolingual; isv matches. |
| | Exits 0 on success, 1 on validation failure. |
| | """ |
| | from pathlib import Path |
| | import os |
| | import sys |
| |
|
| | try: |
| | import pandas as pd |
| | except ImportError: |
| | print("Requires pandas and pyarrow. Install with: pip install -r requirements.txt", file=sys.stderr) |
| | sys.exit(2) |
| |
|
| | REPO_ROOT = Path(__file__).resolve().parent.parent |
| | MONOLINGUAL_PATH = REPO_ROOT / "monolingual" / "isv_sentences.parquet" |
| | PARALLEL_DIR = REPO_ROOT / "parallel" |
| |
|
| | MONO_COLUMNS = {"id", "sentence", "source"} |
| | PARALLEL_COLUMNS = {"id", "isv", "target", "translator", "method", "review_status"} |
| | METHOD_VALUES = {"human", "machine_raw", "machine_postedited"} |
| |
|
| |
|
| | def main() -> None: |
| | errors: list[str] = [] |
| | ci_mode = os.environ.get("GITHUB_ACTIONS", "").lower() in ("true", "1") |
| |
|
| | |
| | df_mono = None |
| | mono_ids: set[str] = set() |
| | mono_id_to_sentence: dict[str, str] = {} |
| |
|
| | if not MONOLINGUAL_PATH.exists(): |
| | errors.append(f"Missing monolingual file: {MONOLINGUAL_PATH}") |
| | else: |
| | df_mono = pd.read_parquet(MONOLINGUAL_PATH) |
| | cols = set(df_mono.columns) |
| | if cols != MONO_COLUMNS: |
| | missing = MONO_COLUMNS - cols |
| | extra = cols - MONO_COLUMNS |
| | if missing: |
| | errors.append(f"monolingual: missing columns: {missing}") |
| | if extra: |
| | errors.append(f"monolingual: unexpected columns: {extra}") |
| | if df_mono["id"].duplicated().any(): |
| | dupes = df_mono[df_mono["id"].duplicated(keep=False)]["id"].unique().tolist() |
| | errors.append(f"monolingual: duplicate ids: {dupes}") |
| | if df_mono["id"].isna().any() or df_mono["sentence"].isna().any() or df_mono["source"].isna().any(): |
| | errors.append("monolingual: null values in id, sentence, or source") |
| | mono_ids = set(df_mono["id"].astype(str)) |
| | mono_id_to_sentence = dict(zip(df_mono["id"].astype(str), df_mono["sentence"].astype(str))) |
| |
|
| | |
| | for path in sorted(PARALLEL_DIR.glob("*.parquet")): |
| | name = path.name |
| | df = pd.read_parquet(path) |
| | cols = set(df.columns) |
| | if cols != PARALLEL_COLUMNS: |
| | missing = PARALLEL_COLUMNS - cols |
| | extra = cols - PARALLEL_COLUMNS |
| | if missing: |
| | errors.append(f"{name}: missing columns: {missing}") |
| | if extra: |
| | errors.append(f"{name}: unexpected columns: {extra}") |
| | if "id" in df.columns and "isv" in df.columns and mono_id_to_sentence: |
| | missing_ids = set(df["id"].astype(str)) - mono_ids |
| | if missing_ids: |
| | errors.append(f"{name}: ids not in monolingual: {list(missing_ids)[:10]}{'...' if len(missing_ids) > 10 else ''}") |
| | |
| | df_ids = df["id"].astype(str) |
| | df_isv = df["isv"].astype(str).str.strip() |
| | expected = df_ids.map(mono_id_to_sentence.get) |
| | in_mono = df_ids.isin(mono_ids) |
| | mismatch = in_mono & (df_isv != expected.str.strip()) |
| | if mismatch.any(): |
| | bad_ids = df.loc[mismatch, "id"].astype(str).tolist() |
| | if ci_mode: |
| | errors.append(f"{name}: id {bad_ids[0]} has isv text that does not match monolingual sentence") |
| | else: |
| | for rid in bad_ids: |
| | errors.append(f"{name}: id {rid} has isv text that does not match monolingual sentence") |
| | if "method" in df.columns and not set(df["method"].dropna().unique()).issubset(METHOD_VALUES): |
| | bad = set(df["method"].dropna().unique()) - METHOD_VALUES |
| | errors.append(f"{name}: invalid method values: {bad}") |
| |
|
| | if errors: |
| | for e in errors: |
| | print(e, file=sys.stderr) |
| | sys.exit(1) |
| | print("Validation passed.") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|