Challenges in Applying Meta-Learning to Foundation Models
Was this section helpful?
Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks, Chelsea Finn, Pieter Abbeel, Sergey Levine, 2017International Conference on Machine Learning (ICML), Vol. 70 (Proceedings of Machine Learning Research (PMLR))DOI: 10.5555/3305890.3306019 - Introduces Model-Agnostic Meta-Learning (MAML), a fundamental algorithm whose computational demands and optimization needs highlight the difficulties of applying meta-learning to large models.
A Survey of Meta-Learning for Few-Shot Learning, Haoqian Wu, Yanzhao Xie, Yunong Wu, Jianqiang Wang, 2021Neurocomputing, Vol. 461 (Elsevier)DOI: 10.1016/j.neucom.2021.08.067 - Provides an overview of meta-learning, discussing various algorithms and their challenges, including the scalability and optimization issues pertinent to foundation models.
LoRA: Low-Rank Adaptation of Large Language Models, Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen, 2021International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.2106.09685 - Introduces Low-Rank Adaptation (LoRA), a parameter-efficient fine-tuning method that addresses computational and memory challenges when adapting large foundation models, offering a practical solution for meta-learning's inner-loop.
On the Optimization of Deep Learning Models for Few-Shot Learning, Maithra Raghu, Shreya Chaudhuri, Maël Obé, J. P. Lewis, Jonathon Shlens, Chelsea Finn, 2020International Conference on Learning Representations (ICLR)DOI: 10.48550/arXiv.1909.02271 - Investigates the optimization landscape and issues in few-shot learning models, offering insights into the stability and hyperparameter sensitivity encountered when scaling these methods.