Prompt Strategies for Short-Term Memory and Context Windows
Was this section helpful?
Attention Is All You Need, Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, 2017arXiv preprint arXiv:1706.03762DOI: 10.48550/arXiv.1706.03762 - Foundational paper introducing the Transformer architecture, which forms the basis for current large language models and their context window mechanism.
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models, Jason Wei, Xuezhi Wang, Dale Schuurmans, Maarten Bosma, Brian Ichter, Fei Xia, Ed Chi, Quoc Le, Denny Zhou, 2022arXiv (arXiv)DOI: 10.48550/arXiv.2201.11903 - Introduces Chain-of-Thought prompting, a method directly relevant to the scratchpad technique for enhancing model reasoning and intermediate thought tracking within the context window.
Lost in the Middle: How Language Models Use Long Contexts, Nelson F. Liu, Kevin Lin, John Hewitt, Ashwin Paranjape, Michele Bevilacqua, Fabio Petroni, Percy Liang, 2023Transactions of the Association for Computational Linguistics (TACL)DOI: 10.48550/arXiv.2307.03172 - Research exploring how large language models process information across long contexts, detailing phenomena such as primacy and recency effects that influence strategic information placement in prompts.
ReAct: Synergizing Reasoning and Acting in Language Models, Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, Yuan Cao, 2023arXiv preprint arXiv:2210.03629DOI: 10.48550/arXiv.2210.03629 - Presents ReAct, an agentic framework combining reasoning and acting through a structured scratchpad approach, demonstrating effective context management for multi-step agent tasks.
Prompt Engineering Guide, Anthropic, 2024 (Anthropic) - Official guide offering practical strategies and best practices for prompt engineering with large language models, including advice on managing context, structuring prompts, and token awareness.