racing-gears / scripts /relabel_clean.py
tobil's picture
v3: deep clean via per-source cross-val + manual curation
4f1fe70 unverified
"""Clean the dataset by training on MNIST (known-clean labels) + Paul Ricard
(visually verified, very clear digits), then using that model to filter/relabel
all racing data.
Strategy:
1. Train a strong CNN on MNIST + Paul Ricard (both have clean, unambiguous labels)
2. Run inference on ALL training data
3. Keep only samples where model agrees with label (or relabel if model is very confident)
4. Do the same for val
5. Rebuild parquet files and composites
Usage:
uv run python scripts/relabel_clean.py
"""
import numpy as np
import pandas as pd
import io
import os
import torch
import torch.nn as nn
import torch.optim as optim
from torch.utils.data import DataLoader, TensorDataset
from PIL import Image
from datasets import Dataset, Image as HFImage
if __name__ != "__main__":
import sys; sys.exit(0)
device = torch.device("cuda" if torch.cuda.is_available()
else "mps" if torch.backends.mps.is_available()
else "cpu")
print(f"Using device: {device}")
class CNN(nn.Module):
def __init__(self):
super().__init__()
self.features = nn.Sequential(
nn.Conv2d(1, 32, 3, padding=1), nn.BatchNorm2d(32), nn.ReLU(), nn.MaxPool2d(2),
nn.Conv2d(32, 64, 3, padding=1), nn.BatchNorm2d(64), nn.ReLU(), nn.MaxPool2d(2),
nn.Conv2d(64, 128, 3, padding=1), nn.BatchNorm2d(128), nn.ReLU(), nn.AdaptiveAvgPool2d(4))
self.classifier = nn.Sequential(
nn.Flatten(), nn.Linear(128 * 16, 256), nn.ReLU(), nn.Dropout(0.4), nn.Linear(256, 10))
def forward(self, x):
return self.classifier(self.features(x))
def load_images(df):
imgs = []
for _, row in df.iterrows():
img = Image.open(io.BytesIO(row["image"]["bytes"])).convert("L")
imgs.append(np.array(img, dtype=np.float32) / 255.0)
return np.stack(imgs)[:, np.newaxis, :, :]
def train_model(X, y, epochs=40):
model = CNN().to(device)
opt = optim.Adam(model.parameters(), lr=1e-3, weight_decay=1e-4)
scheduler = optim.lr_scheduler.CosineAnnealingLR(opt, epochs)
crit = nn.CrossEntropyLoss()
loader = DataLoader(
TensorDataset(torch.tensor(X), torch.tensor(y, dtype=torch.long)),
batch_size=64, shuffle=True)
model.train()
for epoch in range(epochs):
total_loss = 0
for xb, yb in loader:
xb, yb = xb.to(device), yb.to(device)
opt.zero_grad()
loss = crit(model(xb), yb)
loss.backward()
opt.step()
total_loss += loss.item()
scheduler.step()
if (epoch + 1) % 10 == 0:
print(f" epoch {epoch+1}/{epochs} loss={total_loss/len(loader):.4f}")
return model
def predict(model, X):
model.eval()
with torch.no_grad():
probs = torch.softmax(model(torch.tensor(X).to(device)), dim=1).cpu().numpy()
return probs
# --- Step 1: Build clean seed dataset ---
print("\n=== Building clean seed dataset ===")
train_df = pd.read_parquet("data/train-00000-of-00001.parquet")
val_df = pd.read_parquet("data/validation-00000-of-00001.parquet")
# MNIST from train (known clean)
mnist = train_df[train_df["source"] == "mnist"]
print(f"MNIST: {len(mnist)} samples")
# Paul Ricard from train (visually verified, very clear white-on-black)
paul_ricard = train_df[train_df["source"] == "paul-ricard-alpine"]
print(f"Paul Ricard: {len(paul_ricard)} samples")
seed = pd.concat([mnist, paul_ricard], ignore_index=True)
X_seed = load_images(seed)
y_seed = seed["label"].values
print(f"Seed dataset: {len(seed)} samples")
# --- Step 2: Train ensemble on seed ---
print("\n=== Training 5-model ensemble on seed data ===")
models = []
for i in range(5):
print(f"Model {i+1}/5:")
# Bootstrap
idx = np.random.choice(len(X_seed), len(X_seed), replace=True)
model = train_model(X_seed[idx], y_seed[idx])
models.append(model)
# --- Step 3: Predict on all data ---
print("\n=== Predicting on all data ===")
for split_name, df in [("train", train_df), ("validation", val_df)]:
X = load_images(df)
y = df["label"].values
sources = df["source"].values
# Ensemble predict
probs = np.zeros((len(X), 10))
for model in models:
probs += predict(model, X)
probs /= len(models)
preds = probs.argmax(axis=1)
conf = probs.max(axis=1)
# For MNIST: keep as-is (they're the seed)
# For racing: keep only where model agrees OR model is not confident
keep = np.ones(len(df), dtype=bool)
for i in range(len(df)):
if sources[i] == "mnist":
continue # always keep
if preds[i] != y[i] and conf[i] > 0.5:
keep[i] = False
dropped = (~keep).sum()
print(f"\n{split_name}: {len(df)} -> {len(df) - dropped} (dropping {dropped})")
# Per-label breakdown
for label in range(10):
mask = y == label
drop_mask = mask & ~keep
if drop_mask.sum() > 0:
pred_dist = pd.Series(preds[drop_mask]).value_counts().to_dict()
print(f" label={label}: drop {drop_mask.sum()}/{mask.sum()} -> model says {pred_dist}")
df_clean = df[keep].reset_index(drop=True)
print(f" Final distribution:")
print(pd.crosstab(df_clean["label"], df_clean["source"]))
ds = Dataset.from_pandas(df_clean)
ds = ds.cast_column("image", HFImage())
ds.to_parquet(f"data/{split_name}-00000-of-00001.parquet")
print("\nDone! Now run: uv run python scripts/make_composites.py")