Ultrascale Playbook - Expert Parallelism

Notes on training LLMs using expert parallelism

December 13, 2025 · 8 min · 1506 words

Ultrascale Playbook - Context Parallelism

Notes on training LLMs using context parallelism

November 22, 2025 · 11 min · 2284 words

Ultrascale Playbook - Tensor and Sequence Parallelism

Notes on training LLMs using tensor and sequence parallelism

November 11, 2025 · 12 min · 2428 words

Ultrascale Playbook - Pipeline Parallelism

Notes on training LLMs using pipeline parallelism

October 25, 2025 · 14 min · 2921 words

Distributed communication for GPUs (part 2)

Introduction to collective communication operations used for distributed training.

September 13, 2025 · 13 min · 2568 words