Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz Kaiser, and Illia Polosukhin, 2017Advances in Neural Information Processing Systems 30 (NeurIPS)DOI: 10.5555/3295222.3295349 - The seminal paper introducing the Transformer architecture, which forms the basis for most modern large language models, and the specific document used in the practical.
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Narsimha Chilkuri, Michael Grave, Pasquale Minervini, and Sebastian Riedel, 2020Advances in Neural Information Processing Systems 33, Vol. 33 (NeurIPS)DOI: 10.5555/3495289.3495444 - Introduces the Retrieval Augmented Generation (RAG) architecture, which combines parametric and non-parametric memory, significantly enhancing knowledge-intensive natural language processing tasks.
LangChain Python Library Documentation, LangChain Developers, 2024 - The official documentation for the LangChain Python library, offering comprehensive guides, tutorials, and API references for building LLM applications.
OpenAI Embeddings Guide, OpenAI, 2024 (OpenAI) - The official guide from OpenAI explaining the concept of text embeddings and how to use their embedding models, which are central to vector search in RAG systems.