massOfai

Distributed Data Parallel (DDP)

Performance

PyTorch mechanism for synchronous multi-GPU training

What is Distributed Data Parallel (DDP)?

Replicates model across GPUs and synchronizes gradients to keep models consistent.

Real-World Examples

  • PyTorch DDP training across nodes

Related Terms

Learn more about concepts related to Distributed Data Parallel (DDP)