import os, math, glob import datasets from datasets import Features, Value, Array1D from transformers import CLIPProcessor, CLIPModel import torch from PIL import Image from tqdm import tqdm import numpy as np # Optional (recommended for reproducibility) torch.manual_seed(0) # ---------- Config ---------- MODEL_NAME = "openai/clip-vit-base-patch32" BATCH_SIZE = 32 # tune for your machine SHARD_SIZE = 10_000 # write a parquet file every N rows OUT_DIR = "metmuseum_embeddings_streaming" # will contain *.parquet IMG_COL = "jpg" # adjust if column differs (sometimes 'image') ID_COL = "Object ID" # ---------------------------- # 1) Load streaming dataset ds_stream = datasets.load_dataset( "metmuseum/openaccess", split="train", streaming=True ) # 2) Model / processor / device model = CLIPModel.from_pretrained(MODEL_NAME) processor = CLIPProcessor.from_pretrained(MODEL_NAME) device = torch.device("mps" if torch.backends.mps.is_available() else "cpu") model.to(device).eval() # 3) L2 normalize helper def l2_normalize(x, dim=-1, eps=1e-12): return x / (x.norm(p=2, dim=dim, keepdim=True) + eps) # 4) Sharded writer (Parquet via datasets.Dataset) os.makedirs(OUT_DIR, exist_ok=True) shard_idx = 0 rows_in_shard = 0 buffer_ids = [] buffer_vecs = [] emb_dim = None # will set after first batch def flush_shard(): """Write current buffer to a parquet shard and clear it.""" global shard_idx, rows_in_shard, buffer_ids, buffer_vecs, emb_dim if not buffer_ids: return # Ensure emb_dim is known if emb_dim is None: emb_dim = len(buffer_vecs[0]) # Build a small in-memory HF Dataset for this shard with explicit features features = Features({ ID_COL: Value("int32"), "Embedding": Array1D(emb_dim, dtype="float32"), }) shard_ds = datasets.Dataset.from_dict( {ID_COL: buffer_ids, "Embedding": buffer_vecs}, features=features, ) # Write a parquet file (fast & compact) shard_path = os.path.join(OUT_DIR, f"part-{shard_idx:05d}.parquet") shard_ds.to_parquet(shard_path) # Clear buffers / advance shard_idx += 1 rows_in_shard = 0 buffer_ids = [] buffer_vecs = [] # 5) Batch inference loop obj_ids_batch, images_batch = [], [] def flush_batch(): """Run CLIP on the current image batch and append to shard buffer.""" global emb_dim, rows_in_shard, buffer_ids, buffer_vecs if not images_batch: return inputs = processor(images=images_batch, return_tensors="pt") pixel_values = inputs["pixel_values"].to(device) with torch.no_grad(): feats = model.get_image_features(pixel_values=pixel_values) # (B, D) feats = l2_normalize(feats, dim=-1).cpu().numpy().astype("float32") if emb_dim is None: emb_dim = feats.shape[1] # Append to shard buffer buffer_ids.extend([int(x) for x in obj_ids_batch]) buffer_vecs.extend([feats[i] for i in range(feats.shape[0])]) rows_in_shard += feats.shape[0] # Clear batch obj_ids_batch.clear() images_batch.clear() # Iterate stream for item in tqdm(ds_stream, desc="Embedding (streaming)"): oid = item.get(ID_COL) img = item.get(IMG_COL) if oid is None or img is None: continue # Ensure PIL RGB if isinstance(img, Image.Image): pil_img = img.convert("RGB") else: try: pil_img = Image.fromarray(img).convert("RGB") except Exception: continue obj_ids_batch.append(oid) images_batch.append(pil_img) if len(images_batch) >= BATCH_SIZE: flush_batch() if rows_in_shard >= SHARD_SIZE: flush_shard() # Flush remainder flush_batch() flush_shard() print(f"Wrote {shard_idx} shard(s) to {OUT_DIR}")