Mixed Precision Training, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1710.03740 - This foundational paper introduces the techniques for training deep neural networks with mixed precision, primarily using FP16, demonstrating significant speedups and memory savings.
Automatic Mixed Precision (AMP), PyTorch Documentation, 2025 (PyTorch Foundation) - Official documentation explaining how to use automatic mixed precision in PyTorch, covering practical implementation details and best practices for memory and speed benefits.
NVIDIA Deep Learning Performance Guide, NVIDIA Corporation, 2023 - This guide provides recommendations and explanations for optimizing deep learning performance on NVIDIA GPUs, including the role of mixed precision and Tensor Cores.