ReAct: Synergizing Reasoning and Acting in Language Models, Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik Narasimhan, Yuan Cao, 2022International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2210.03629 - Introduces the ReAct paradigm, a foundational approach for LLM agents combining reasoning and acting, including handling tool interactions and their potential outcomes. This approach is valuable for graceful error recovery in agentic workflows.
Reflexion: Language Agents with Verbal Reinforcement Learning, Noah Shinn, Federico Cassano, Edward Berman, Ashwin Gopinath, Karthik Narasimhan, Shunyu Yao, 2023arXiv preprint arXiv:2303.11366DOI: 10.48550/arXiv.2303.11366 - Presents a framework for LLM agents to perform verbal reinforcement learning by reflecting on past action sequences and generating self-reflection prompts to guide future actions. This is directly relevant to prompting for self-critique and iterative modification upon failure.
Constitutional AI: Harmlessness from AI Feedback, Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Anna Chen, Andy Jones, Anna Goldieberger, Aziza Mirrashed, Cameron McKinnon, Carol Chen, Catherine Olsson, Chris Conly, David Drain, Dawn Drain, Deep Ganguli, Dustin Li, Eli Tran-Johnson, Ethan Perez, Jackson Kernion, Jasmine Sittler, Jennifer Glowinski, Jeremy Scheurer, Jessica Kerr, Josh Jacobson, Kristen Lee, Liane Lovitt, Lisa Wang, Michael Sellitto, Mo Mukherjee, Nicholas Joseph, Noemi Mercado, Nova DasSarma, Robert Lasenby, Robin Larson, Sam Ringer, Shauna Gordon-McKeon, Simon Lefevre, Tristan Hume, Zac Hatfield-Dodds, Danny Hernandez, Daniela Amodei, Dario Amodei, Jack Clark, Sam McCandlish, Tom Brown, Jared Kaplan, 2022arXiv preprint arXiv:2209.07858DOI: 10.48550/arXiv.2209.07858 - While primarily focused on safety, this paper introduces a method where AI models critique and revise their own outputs based on a set of principles, showing a robust form of AI-driven self-assessment and correction.