Long Short-Term Memory, Sepp Hochreiter, Jürgen Schmidhuber, 1997Neural Computation, Vol. 9 (MIT Press)DOI: 10.1162/neco.1997.9.8.1735 - This original paper introduces the Long Short-Term Memory (LSTM) network, a core component for building recurrent neural networks, including stacked architectures.
Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - This textbook offers a foundational background in deep learning, covering recurrent neural networks, LSTMs, GRUs, and principles of deep architectures. Chapters 10 and 11 are particularly relevant.
tf.keras.layers.LSTM, TensorFlow Developers, 2024 - The official documentation for the Keras LSTM layer, including details on the return_sequences parameter, which is important for stacking recurrent layers in TensorFlow/Keras models.
torch.nn.LSTM, PyTorch Developers, 2024 (PyTorch Foundation) - The official documentation for the PyTorch LSTM module, explaining parameters such as num_layers and the structure of output sequences, which is relevant for building stacked RNNs in PyTorch.