AI & ML interests

None defined yet.

Recent Activity

Yuwei-Niu  updated a Space about 7 hours ago
LMMs-Lab-Speedrun/README
Yuwei-Niu  published a Space about 7 hours ago
LMMs-Lab-Speedrun/README
pufanyi  updated a dataset about 12 hours ago
LMMs-Lab-Speedrun/Data_NanoVLM
View all activity

Organization Card

NanoVLM Speedrun

The most striking thing about the modded-nanogpt experiments is that they expose how much of deep learning is just bloat. To apply this to Vision-Language Models (VLMs), you have to stop acting like a researcher and start acting like a hacker. You aren't trying to follow academic standards; you are trying to maximize the movement of bits through silicon. We introduce NanoVLM Speedrun: a minimalist VLM recipe designed to strip away the bloat. We provide the bare-minimum components required to bridge the training and evaluation pipeline, enabling lightning-fast iteration and reproduction.

The Recipe (2026H1)

  • LLM: Qwen/Qwen3-0.6B
  • Vision Encoder: google/siglip2-so400m-patch16-naflex
  • Projector: Classic LLaVA-style 2-layer MLP
  • Training Paradigm: A streamlined two-stage approach:
    • Stage 1: Projector-only alignment (tuning the projector between vision and language).
    • Stage 2: End-to-end instruction tuning (tuning both the projector and the LLM).

Data Preparation

We utilize the curated LMMs-Lab-Speedrun/Data_NanoVLM collection.

For more information about training, please refer to NanoVLM Speedrun.