Mixed Precision
Neural Networks
Using lower precision where safe to speed training/inference
What is Mixed Precision?
Combines FP16 and FP32 to reduce memory and increase throughput, while preserving model accuracy with careful scaling.
Real-World Examples
- •NVIDIA Apex, PyTorch native AMP
Related Terms
Learn more about concepts related to Mixed Precision