Efficient Estimation of Word Representations in Vector Space, Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean, 2013arXiv preprint arXiv:1301.3781 - This paper introduces the Word2Vec model, which significantly advanced the field of learning dense word embeddings that capture semantic and syntactic relationships.
GloVe: Global Vectors for Word Representation, Jeffrey Pennington, Richard Socher, Christopher Manning, 2014Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) (Association for Computational Linguistics)DOI: 10.3115/v1/D14-1162 - Presents GloVe, another influential model for generating word embeddings by factoring a global word-word co-occurrence matrix, offering an alternative perspective to Word2Vec.
Deep Learning, Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016 (MIT Press) - A comprehensive textbook covering fundamental concepts of deep learning, including detailed explanations of neural network architectures, representation learning, and word embeddings.
tf.keras.layers.Embedding, TensorFlow Developers, 2024 - Official documentation for TensorFlow's Keras Embedding layer, describing its usage, parameters, and how it functions as a trainable lookup table for token embeddings.