File size: 6,929 Bytes
a473570
 
 
 
 
 
 
 
 
 
65ea356
 
 
a473570
 
 
 
21561fa
a473570
21561fa
a473570
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
911e14b
 
a473570
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
---
license: mit
language:
  - en
tags:
  - optimization
  - muon
  - deep-learning
  - pytorch
  - distributed-computing
  - ZeRO
  - tensor parallelism
  - data parallelism
  - tutorial
  - cpu-friendly
---

# 🧩 The "Muon is Scalable" Blueprint: A Distributed Muon Engineering Breakdown(CPU-Friendly, tutorial style)

A practical, **annotated engineering breakdown** for the Muon optimizer — extended into a fully distributed (DP × TP) version that actually runs on **plain CPU/Gloo**, so broke-but-curious builders can still get their hands dirty.

This is the expert-level systems engineering companion to my original [**Understanding the Muon Optimizer**](https://huggingface.co/datasets/bird-of-paradise/muon-tutorial) tutorial.

> 💡 _“Because sometimes the best way to learn the distributed nightmare is to get your hands dirty and your eyes crossed.”_ 🤪

---

## 🌕 Why This Exists

Most public Muon examples (like MoonShot’s PoC) are designed for multi-GPU NCCL clusters, making them impossible to run or debug for most of us. In addition, most documentation for distributed systems is written by experts, for experts, making it a "nightmare" to learn. My goal is to change that.

This repository is **not** a "simplified" version that "flattens the depth" of the work.

Instead, it's a **didactic refactor**. I've taken the complex, real-world PoC and optimized it for *readability* and *learning*, so you can see the "blueprint" behind the "chaos":
-   Fully **annotated** to **demonstrate** data parallel (ZeRO-1) + tensor parallel (TP) orchestration end-to-end.
-   Understand every “distributed acrobatic” step (`DP gather``TP gather``Newton–Schulz``TP shard``DP shard`).
-   The code is optimized to highlight **symmetrical logic** and **consistent naming**, showing the "opposite arrow" data flow of the "virtual map" (`dist_meta`).
-   It's built to run **on a single CPU machine or Colab notebook**.

---

## 🧠 The “Turtle Speed” Breakthrough: The Full Story

This code is complex. It's a "distributed nightmare" 🫠.

Instead of a traditional, long-form `README`, the best documentation is the "making of" story. I've chronicled my entire journey of reverse-engineering and debugging this code in my "Turtle Speed Breakthrough" series on Medium.

* **[Part 1: The “Turtle Speed” Breakthrough: Decoding Distributed Optimizers](https://medium.com/@jenwei0312/the-turtle-speed-breakthrough-decoding-distributed-optimizers-from-fsdp-to-muons-secret-sauce-64fc76f20cd7)**
* **[Part 2: My Map of the Distributed Nightmare (The Blueprint)](https://medium.com/@jenwei0312/the-turtle-speed-breakthrough-part-2-the-blueprint-for-distributed-chaos-37fe343e7aa9)**
* **[Part 3: The Final Bugs and "Aha!" Moments](https://medium.com/@jenwei0312/the-turtle-speed-breakthrough-part-3-my-map-of-the-distributed-nightmare-b10ff4affd56)**

This tutorial is the final, runnable code that resulted from that deep dive.

---

## 🚀 Quick Start

Run the CPU-safe, fully-annotated notebook right from your browser:

[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](<distributed_muon_cpu.ipynb>)

Or, you can clone this repo and run the Python script locally to simulate an 8-process cluster on your CPU:

```bash
git clone https://huggingface.co/datasets/bird-of-paradise/muon-distributed
cd muon-distributed

# This will spawn 8 processes and run the full test
!python distributed_muon_cpu.py
````

(Note: For the original, un-annotated, buggy Moonshot PoC that this work is based on, you can find it in this [commit](https://github.com/NVIDIA/Megatron-LM/pull/1428/commits/f432fbe45c169aeb5a0805ff6f41e13f989c6730#diff-8fe91f4096ff232fc6f97b17e60e619eda92b6dffc80b4573a23e06aa56d2559).)

-----

## 🗂️ What's Inside (File Guide)

  * `distributed_muon_cpu.ipynb`: **(Start Here)** The Colab-friendly notebook that walks through the environment fixes and runs the code.
  * `distributed_muon_cpu.py`: The final, **tested, fixed, and heavily-annotated** Python script. This is the "golden" code that runs on a CPU-only environment using the `"gloo"` backend.
  * `distributed_muon.py`: My **annotated and logically debugged** version of the *GPU* code. This is for users who have a multi-GPU `"nccl"` environment. (Note: Since I don't have a multi-GPU cluster, this version is untested... unless someone wants to sponsor me with some GPUs! 😉)

-----

## 🎓 What You'll Learn (The "Nightmare" Blueprint)

By exploring this code, you'll see the *real* implementation of the concepts I discuss in my articles:

  * **The 2D Grid:** How to set up orthogonal `dist_group` (DP) and `tp_group` (TP) handles.
  * **The "Aisles" & "Pallets":** How `param_groups` (`buffer_idx`) and communication `buckets` (`bucket_idx`) are used to organize parameters.
  * **The "Virtual Buffer":** How a "master plan" (`dist_meta` and `global_buffer_size`) is used to manage memory for sharding (ZeRO-1).
  * **The Acrobatic Data Flow:** The full `(DP gather -> TP gather) -> (Run Math) -> (TP shard -> DP shard)` journey.
  * **The Nuance:** You'll see *why* we bucket the slow DP `all_gather` but *don't* need to bucket the fast, on-node TP `all_gather`.

-----

## 🐞 Summary of All Fixes

This repo isn't just a copy-paste. It's the result of a week-long debugging "nightmare." Here are all the bugs we had to find and fix to make it run:

| Issue | Problem | Solution |
| :--- | :--- | :--- |
| **Logic Bug \#1** | Missing `params = group["params"]` | Added the line in the Muon update loop. |
| **Logic Bug \#2** | `ns_input` was 1D after unpacking, crashing `zeropower`.| Changed `.view(-1)` to `.view(dist_meta.shape)` to restore the 2D shape. |
| **Env Bug \#1** | Hardcoded `"nccl"` backend. | Changed `dist.init_process_group` to use `"gloo"`. |
| **Env Bug \#2** | Hardcoded `'cuda'` device. | Changed `gen_param_and_grads` to use `'cpu'`. |
| **Env Bug \#3** | `mp.spawn()` crashes in Jupyter/Colab. | The `.ipynb` runs the code as a `!python` subprocess, bypassing the notebook kernel. |

-----

## 📖 Citation

If you use this tutorial in your work, please cite the original Muon paper and this tutorial.

```bibtex
@misc{wei2025muondistributed,
  author = {Wei, Jen},
  title = {A CPU-Friendly Tutorial for Distributed Muon (DPxTP)},
  year = {2025},
  howpublished = {\url{[https://huggingface.co/datasets/](https://huggingface.co/datasets/)<your-hf-handle>/muon-distributed}}
}

@misc{jordan2024muon,
  author = {Jordan, Keller, et al.},
  title = {Muon: An optimizer for hidden layers in neural networks},
  year = {2024},
  url = {[https://kellerjordan.github.io/posts/muon/](https://kellerjordan.github.io/posts/muon/)}
}

@misc{liu2025muonscalable,
  author = {Liu, Jingyuan, et al.},
  title = {Muon is Scalable for LLM Training},
  year = {2025},
  url = {[https://arxiv.org/abs/2502.16982](https://arxiv.org/abs/2502.16982)}
}
```