Long Short-Term Memory, Sepp Hochreiter, Jürgen Schmidhuber, 1997Neural Computation, Vol. 9 (MIT Press)DOI: 10.1162/neco.1997.9.8.1735 - This is the seminal paper introducing the Long Short-Term Memory (LSTM) network architecture, which is a core component mentioned in the section for capturing temporal dependencies in time series data.
Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - Chapter 10 of this authoritative textbook provides a comprehensive theoretical background on Recurrent Neural Networks, including LSTMs and GRUs, and their applications in sequence modeling and time series.
Time series forecasting, TensorFlow Developers, 2024 (Google) - This official TensorFlow tutorial provides practical examples and code for applying RNNs (LSTMs and GRUs) to time series forecasting, covering essential steps like data preprocessing, windowing, and implementing different model architectures in Keras.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation, Kyunghyun Cho, Bart van Merrienboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and Yoshua Bengio, 2014Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)DOI: 10.48550/arXiv.1406.1078 - This paper introduces the Gated Recurrent Unit (GRU) as a simpler alternative to LSTM, which is also widely used for sequence modeling and time series forecasting due to its ability to capture long-term dependencies.
Deep learning for time series forecasting: A survey, Bryan Lim and Stefan Zohren, 2021European Journal of Operational Research, Vol. 296 (Elsevier)DOI: 10.1016/j.ejor.2020.08.006 - This comprehensive survey provides an overview of recent advancements in deep learning models for time series forecasting, discussing various architectures, preprocessing techniques, and their applications, offering a broader context for the methods presented in the section.