Fine-tune a pretrained model with the SFTTrainer, Hugging Face, 2024 (Hugging Face) - Official documentation for the SFTTrainer in the Hugging Face trl library, providing practical guidance on data preparation, special tokens, and loss masking for supervised fine-tuning.
Finetuned Language Models Are Zero-Shot Learners, Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, Quoc V. Le, 2022arXivDOI: 10.48550/arXiv.2109.01652 - Introduced instruction tuning, a core concept for SFT, demonstrating how diverse task formats (prompts and completions) can be used to improve model generalization and zero-shot performance.