Automatic Mixed Precision (AMP), PyTorch Team, 2025 (PyTorch Documentation) - Official documentation for PyTorch's Automatic Mixed Precision (AMP) feature, including usage of torch.autocast and GradScaler with bfloat16.
Mixed-Precision Training, NVIDIA Developer Documentation, 2024 (NVIDIA) - NVIDIA's guide on mixed-precision training techniques, including details on FP16 and BF16, hardware support, and performance considerations for NVIDIA GPUs.
Mixed-Precision Training of Deep Neural Networks, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1710.03740 - Introduces effective techniques like loss scaling for stable FP16 training, providing foundational context for why BF16's wider range is beneficial.