rirmega / scripts /make_checksums.py
mandipgoswami's picture
Upload 21 files
ee24db9 verified
# scripts/make_checksums.py
import argparse, csv, hashlib, pathlib, sys, os
from collections import defaultdict
def sha256(p, buf=1024*1024):
h = hashlib.sha256()
with open(p, "rb") as f:
while chunk := f.read(buf):
h.update(chunk)
return h.hexdigest()
def norm_header(h: str) -> str:
return (h or "").strip().lstrip("\ufeff").lower().replace(" ", "_")
def norm_path_for_match(p: str) -> str:
# Normalize separators and remove leading .\ or ./, make POSIX-ish
q = p.replace("\\", "/")
while q.startswith("./") or q.startswith(".\\"):
q = q[2:]
return q
def main():
ap = argparse.ArgumentParser()
ap.add_argument("--meta", default="data/metadata/metadata.csv",
help="Path to metadata.csv")
ap.add_argument("--audio-root", default="data/audio",
help="Root folder containing audio files")
ap.add_argument("--path-col", default=None,
help="Column name that holds paths (default: auto-detect; prefers 'file_path')")
ap.add_argument("--checksum-col", default="checksum_sha256",
help="Column name to write sha256 into")
args = ap.parse_args()
root = pathlib.Path(__file__).resolve().parents[1]
audio_root = (root / args.audio_root).resolve()
meta_path = (root / args.meta).resolve()
if not meta_path.exists():
sys.exit(f"[ERR] metadata file not found: {meta_path}")
if not audio_root.exists():
print(f"[WARN] audio root not found yet: {audio_root}")
# Build map: relative path (POSIX) -> sha256
print("[INFO] Scanning audio files for checksums…")
hashmap = {}
for p in audio_root.rglob("*"):
if p.suffix.lower() not in {".wav", ".flac"} or not p.is_file():
continue
rel = p.relative_to(root).as_posix()
hashmap[norm_path_for_match(rel)] = sha256(p)
# Read CSV with dialect sniffing and normalized headers
print(f"[INFO] Reading metadata: {meta_path}")
raw = meta_path.read_text(encoding="utf-8", errors="replace")
try:
dialect = csv.Sniffer().sniff(raw.splitlines()[0] if raw else ",")
except Exception:
dialect = csv.excel # fallback to comma
rows = []
with open(meta_path, newline="", encoding="utf-8", errors="replace") as f:
reader = csv.reader(f, dialect)
try:
headers = next(reader)
except StopIteration:
sys.exit("[ERR] metadata.csv appears empty.")
norm_headers = [norm_header(h) for h in headers]
hdr_map = {norm_header(h): i for i, h in enumerate(headers)}
# Choose the path column
candidate_names = [norm_header(args.path_col)] if args.path_col else [
"file_path", "filepath", "path", "relative_path", "audio_path", "wav", "rir_path"
]
path_col_norm = next((c for c in candidate_names if c in hdr_map), None)
if not path_col_norm:
msg = (f"[ERR] Could not find a path column. Looked for any of: "
f"{candidate_names}. Available columns: {norm_headers}")
sys.exit(msg)
checksum_col_norm = norm_header(args.checksum_col)
# If checksum column absent, append it
if checksum_col_norm not in hdr_map:
headers.append(args.checksum_col)
norm_headers.append(checksum_col_norm)
checksum_idx = len(headers) - 1
else:
checksum_idx = hdr_map[checksum_col_norm]
path_idx = hdr_map[path_col_norm]
# Process rows
rows.append(headers) # header row for writing back
for i, row in enumerate(reader, start=1):
# pad short rows
if len(row) < len(headers):
row += [""] * (len(headers) - len(row))
# Normalize the path for lookup
csv_path_raw = (row[path_idx] or "").strip()
if not csv_path_raw:
print(f"[WARN] row {i}: empty path cell; leaving checksum blank")
rows.append(row)
continue
# Try multiple lookup strategies
candidates = []
# 1) CSV path as given (normalized)
candidates.append(norm_path_for_match(csv_path_raw))
# 2) If CSV path is absolute, try making it relative to project root
p = pathlib.Path(csv_path_raw)
if p.is_absolute():
try:
rel = p.relative_to(root).as_posix()
candidates.append(norm_path_for_match(rel))
except Exception:
pass
# 3) If CSV path is relative to audio_root
try:
rel2 = (audio_root / csv_path_raw).resolve().relative_to(root).as_posix()
candidates.append(norm_path_for_match(rel2))
except Exception:
pass
# 4) Fallback: match by basename if unique
basename = pathlib.Path(csv_path_raw).name
if basename:
# build once a reverse index by basename
pass
# deduplicate candidates
candidates = list(dict.fromkeys(candidates))
sha = ""
for cand in candidates:
sha = hashmap.get(cand, "")
if sha:
break
# As a last resort, basename matching (unique)
if not sha and basename:
matches = [v for k, v in hashmap.items() if pathlib.Path(k).name == basename]
if len(matches) == 1:
sha = matches[0]
row[checksum_idx] = sha
if not sha:
print(f"[WARN] row {i}: no match for '{csv_path_raw}' (tried {len(candidates)} candidates)")
rows.append(row)
# Write back CSV (same dialect; UTF-8)
print(f"[INFO] Writing updated metadata with checksums → {meta_path}")
with open(meta_path, "w", newline="", encoding="utf-8") as f:
writer = csv.writer(f, dialect)
writer.writerows(rows)
print("[DONE] Checksums inserted. "
f"Found hashes for ~{sum(1 for r in rows[1:] if r[checksum_idx])} rows.")
if __name__ == "__main__":
main()