Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - A foundational textbook offering comprehensive coverage of deep learning principles, including optimization algorithms, gradient descent, and the rationale for dynamic learning rate adjustments.
SGDR: Stochastic Gradient Descent with Warm Restarts, Ilya Loshchilov, Frank Hutter, 2017International Conference on Learning RepresentationsDOI: 10.48550/arXiv.1608.03983 - Introduces the concept of cosine annealing with warm restarts, a technique for learning rate scheduling that frequently achieves superior performance in deep learning training.
How to adjust learning rate, PyTorch Documentation, 2023 (PyTorch) - Official documentation providing practical guidance and examples for implementing various learning rate schedulers within the PyTorch deep learning framework.