Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks, Patrick Lewis, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, Mike Lewis, Wen-tau Yih, Tim Rocktäschel, Sebastian Riedel, Douwe Kiela, 2020NeurIPS 2020DOI: 10.48550/arXiv.2005.11401 - This paper introduced the Retrieval-Augmented Generation (RAG) paradigm, explaining its architecture and benefits.
A Survey on Approximate Nearest Neighbor Search, Jianqiu Xu, Yang Cao, Yulong Huang, Bin Wang, Xinjie Zhang, Ruihui Zhao, 2020Big Data Mining and Analytics, Vol. 3 (Tsinghua University Press)DOI: 10.26599/BDMA.2020.1030005 - Discusses various techniques and algorithms for approximate nearest neighbor search, which is fundamental to the vector search component of RAG.
Language Models are Few-Shot Learners, Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei, 2020Advances in Neural Information Processing Systems (NeurIPS) 33, Vol. 33DOI: 10.48550/arXiv.2005.14165 - This paper introduced in-context learning, demonstrating how large language models can perform new tasks given a few examples, setting groundwork for prompt engineering.
LangChain Documentation, Harrison Chase, 2024 (LangChain) - Provides comprehensive guides and examples for building LLM applications, including detailed implementation patterns for RAG pipelines.