README / README.md
Yuwei-Niu's picture
Update README.md
7b3952b verified
metadata
title: README
emoji: 
colorFrom: purple
colorTo: pink
sdk: static
pinned: false

NanoVLM Speedrun

The most striking thing about the modded-nanogpt experiments is that they expose how much of deep learning is just bloat. To apply this to Vision-Language Models (VLMs), you have to stop acting like a researcher and start acting like a hacker. You aren't trying to follow academic standards; you are trying to maximize the movement of bits through silicon. We introduce NanoVLM Speedrun: a minimalist VLM recipe designed to strip away the bloat. We provide the bare-minimum components required to bridge the training and evaluation pipeline, enabling lightning-fast iteration and reproduction.

The Recipe (2026H1)

  • LLM: Qwen/Qwen3-0.6B
  • Vision Encoder: google/siglip2-so400m-patch16-naflex
  • Projector: Classic LLaVA-style 2-layer MLP
  • Training Paradigm: A streamlined two-stage approach:
    • Stage 1: Projector-only alignment (tuning the projector between vision and language).
    • Stage 2: End-to-end instruction tuning (tuning both the projector and the LLM).

Data Preparation

We utilize the curated LMMs-Lab-Speedrun/Data_NanoVLM collection.

For more information about training, please refer to NanoVLM Speedrun.