Neural Machine Translation by Jointly Learning to Align and Translate, Dzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio, 2015International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1409.0473 - This pioneering paper introduces the attention mechanism in sequence-to-sequence models, detailing how a context vector is created as a weighted sum of encoder hidden states to focus on relevant input parts.
Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, 2017Advances in Neural Information Processing Systems 30 (NeurIPS 2017)DOI: 10.48550/arXiv.1706.03762 - This foundational paper introduces the Transformer architecture, which entirely relies on attention mechanisms and explicitly defines the computation of context vectors via weighted sums of Value vectors.