"""Iterative dataset cleaning: cross-val clean train 3x, then ensemble-clean val.""" import numpy as np import pandas as pd import io import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader, TensorDataset from PIL import Image from sklearn.model_selection import StratifiedKFold from datasets import Dataset, Image as HFImage if __name__ != "__main__": import sys; sys.exit(0) class SmallCNN(nn.Module): def __init__(self): super().__init__() self.features = nn.Sequential( nn.Conv2d(1, 32, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(32, 64, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(64, 128, 3, padding=1), nn.ReLU(), nn.AdaptiveAvgPool2d(4)) self.classifier = nn.Sequential( nn.Flatten(), nn.Linear(128 * 16, 256), nn.ReLU(), nn.Dropout(0.3), nn.Linear(256, 10)) def forward(self, x): return self.classifier(self.features(x)) def load_images(df): imgs = [] for _, row in df.iterrows(): img = Image.open(io.BytesIO(row["image"]["bytes"])).convert("L") imgs.append(np.array(img, dtype=np.float32) / 255.0) return np.stack(imgs)[:, np.newaxis, :, :] def crossval_predict(X, y, n_splits=5, epochs=30): pred_probs = np.zeros((len(X), 10)) skf = StratifiedKFold(n_splits=n_splits, shuffle=True, random_state=42) for fold, (tr, va) in enumerate(skf.split(X, y)): print(f" fold {fold + 1}/{n_splits}...", flush=True) model = SmallCNN() opt = optim.Adam(model.parameters(), lr=1e-3) crit = nn.CrossEntropyLoss() loader = DataLoader( TensorDataset(torch.tensor(X[tr]), torch.tensor(y[tr], dtype=torch.long)), batch_size=64, shuffle=True) model.train() for _ in range(epochs): for xb, yb in loader: opt.zero_grad() crit(model(xb), yb).backward() opt.step() model.eval() with torch.no_grad(): pred_probs[va] = torch.softmax(model(torch.tensor(X[va])), dim=1).numpy() return pred_probs def clean_round(df, round_num, conf_threshold=0.6): print(f"\n=== Round {round_num}: cross-val clean ({len(df)} samples) ===", flush=True) X = load_images(df) y = df["label"].values pred_probs = crossval_predict(X, y) preds = pred_probs.argmax(axis=1) acc = (preds == y).mean() print(f" OOF accuracy: {acc:.3f}") conf = pred_probs.max(axis=1) bad = (preds != y) & (conf > conf_threshold) print(f" Removing {bad.sum()} samples (conf > {conf_threshold})") for label in range(10): mask = (y == label) & bad if mask.sum() > 0: pred_dist = pd.Series(preds[mask]).value_counts().to_dict() print(f" label={label}: drop {mask.sum()} -> model says {pred_dist}") return df[~bad].reset_index(drop=True), acc # === Iterative train cleaning === train_df = pd.read_parquet("data/train-00000-of-00001.parquet") for round_num in range(1, 4): train_df, acc = clean_round(train_df, round_num) if acc > 0.97: print(" Accuracy high enough, stopping early") break print(f"\nFinal train: {len(train_df)}") print(pd.crosstab(train_df["label"], train_df["source"])) # === Ensemble clean val === print(f"\n=== Cleaning val with 7-model ensemble ===", flush=True) X_train = load_images(train_df) y_train = train_df["label"].values val_df = pd.read_parquet("data/validation-00000-of-00001.parquet") X_val = load_images(val_df) y_val = val_df["label"].values val_probs = np.zeros((len(X_val), 10)) for i in range(7): print(f" model {i + 1}/7...", flush=True) model = SmallCNN() opt = optim.Adam(model.parameters(), lr=1e-3) crit = nn.CrossEntropyLoss() idx = np.random.choice(len(X_train), len(X_train), replace=True) loader = DataLoader( TensorDataset(torch.tensor(X_train[idx]), torch.tensor(y_train[idx], dtype=torch.long)), batch_size=64, shuffle=True) model.train() for _ in range(30): for xb, yb in loader: opt.zero_grad() crit(model(xb), yb).backward() opt.step() model.eval() with torch.no_grad(): val_probs += torch.softmax(model(torch.tensor(X_val)), dim=1).numpy() val_probs /= 7 val_pred = val_probs.argmax(axis=1) val_conf = val_probs.max(axis=1) print(f" Val accuracy vs labels: {(val_pred == y_val).mean():.3f}") bad_val = (val_pred != y_val) & (val_conf > 0.7) print(f" Removing {bad_val.sum()} val samples") for label in range(10): mask = (y_val == label) & bad_val if mask.sum() > 0: pred_dist = pd.Series(val_pred[mask]).value_counts().to_dict() print(f" label={label}: drop {mask.sum()} -> model says {pred_dist}") val_final = val_df[~bad_val].reset_index(drop=True) print(f"\nFinal val: {len(val_final)}") print(pd.crosstab(val_final["label"], val_final["source"])) # === Write === for name, df in [("train", train_df), ("validation", val_final)]: ds = Dataset.from_pandas(df.reset_index(drop=True)) ds = ds.cast_column("image", HFImage()) ds.to_parquet(f"data/{name}-00000-of-00001.parquet") print("\nDone!")