Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - This book is a comprehensive reference for deep learning, with dedicated chapters explaining recurrent neural networks (Chapter 10) and their applications, which forms the basis for sequence classification.
CS224n: Natural Language Processing with Deep Learning, Christopher Manning, Abigail See, 2023Online Course Materials (Stanford University) - This course extensively covers recurrent neural networks, LSTMs, GRUs, and their application to various NLP tasks, including sequence classification for sentiment analysis and other problems, often detailing architectural considerations.
Long Short-Term Memory, Sepp Hochreiter and Jürgen Schmidhuber, 1997Neural Computation, Vol. 9 (MIT Press)DOI: 10.1162/neco.1997.9.8.1735 - This is the original paper introducing the Long Short-Term Memory (LSTM) network architecture, which is fundamental to modern recurrent neural network applications, including sequence classification.
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation, Kyunghyun Cho, Bart van Merriënboer, Çağlar Gülçehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, Yoshua Bengio, 2014Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Association for Computational Linguistics)DOI: 10.3115/v1/D14-1179 - This paper introduces Gated Recurrent Units (GRUs) as a simplification of LSTMs, which are widely used in sequence modeling tasks, including classification, for their efficiency and performance.