ylecun/mnist
Viewer • Updated • 70k • 82.5k • 239
This is a fully connected MNIST digit classifier trained as part of the
dlab deep-learning experimentation roadmap.
The model is intentionally an MLP rather than a CNN. It is useful as a strong baseline for studying optimization, regularization, augmentation, seed variance, and the ceiling of non-convolutional models on MNIST.
10-seed confirmation sweep:
| metric | result |
|---|---|
| validation accuracy | 99.3600% ± 0.0817 pp |
| validation loss | 0.15172 ± 0.00235 |
| test accuracy | 99.4470% ± 0.0195 pp |
| test loss | 0.14746 ± 0.00034 |
| test errors | 55.3 ± 1.95 / 10000 |
model.ckpt: PyTorch Lightning checkpoint from the best run.model.onnx: ONNX export of the EMA/current model state used for evaluation.config.yaml: resolved Hydra training config.metrics.csv: training metrics from the run.metadata.json: compact metadata for inference and provenance.Inputs should be MNIST grayscale images converted to tensors and normalized with:
mean = [0.1307]
std = [0.3081]
The ONNX model expects float tensors shaped [batch, 1, 28, 28] under the input
name images, and returns class logits under the output name logits.