Ultrascale Playbook - Pipeline Parallelism
Notes on training LLMs using pipeline parallelism
Notes on training LLMs using pipeline parallelism
My talk on training LLMs at Pydata MCR
Introduction to collective communication operations used for distributed training.
Introduction to distributed communication for GPUs.
Notes on choosing appropriate batch size and compute for training LLMs