Hugging Face Transformers Library Documentation, Hugging Face team, 2024 - Official and comprehensive documentation for the Hugging Face transformers library, essential for practical implementation and understanding its full capabilities.
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova, 2018Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Vol. 1DOI: 10.48550/arXiv.1810.04805 - Introduces the BERT model, a foundational pre-trained Transformer that revolutionized NLP and demonstrated the power of the pre-training and fine-tuning paradigm, which is central to using model libraries.