Mixed-Precision Training for Deep Neural Networks, Paulius Micikevicius, Sharan Narang, Jonah Alben, Gregory Diamos, Erich Elsen, David Garcia, Boris Ginsburg, Michael Houston, Oleksii Kuchaiev, Ganesh Venkatesh, Hao Wu, 2018International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1710.03740 - Foundational paper introducing mixed-precision training for deep neural networks, including the use of FP16, loss scaling to prevent underflow, and benefits on NVIDIA GPUs with Tensor Cores.
Automatic Mixed Precision for torch.cuda operations, PyTorch Authors, 2025 - Official PyTorch documentation explaining how to enable and use Automatic Mixed Precision (AMP) through torch.cuda.amp.autocast and torch.cuda.amp.GradScaler, providing practical implementation details.
Mixed precision training, TensorFlow Authors, 2023 - Official TensorFlow guide on how to implement mixed precision training in Keras using tf.keras.mixed_precision.Policy, detailing the setup and usage for performance improvements.