LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2106.09685 - Introduces Low-Rank Adaptation (LoRA), a widely adopted technique for efficiently fine-tuning large language models.
Parameter-Efficient Transfer Learning for NLP, Neil Houlsby, Andrei Giurgiu, Stanislaw Jastrzebski, Bruna Morrone, Quentin de Laroussilhe, Andrea Gesmundo, Mona Attariyan, Sylvain Gelly, 2019arXiv preprintDOI: 10.48550/arXiv.1902.00751 - Presents adapter modules, an early and influential method for parameter-efficient fine-tuning by injecting small, trainable layers.
PEFT: Parameter-Efficient Fine-tuning, Hugging Face, 2024 (Hugging Face) - Official documentation for the Hugging Face PEFT library, offering practical guidance and implementation details for various PEFT methods.