A paradigm enabling models to learn how to learn across tasks efficiently.
Meta-learning, often described as "learning to learn," is a machine learning paradigm in which a model is trained not just to solve a specific task, but to improve its own learning process across a distribution of tasks. Rather than starting from scratch on each new problem, a meta-learned model leverages prior experience to adapt rapidly—often from just a handful of examples. This makes meta-learning especially valuable in low-data regimes, where traditional deep learning approaches struggle due to their dependence on large labeled datasets.
The mechanics of meta-learning generally involve two nested loops of optimization. The outer loop trains the model across many different tasks, updating a set of shared parameters or a learning strategy. The inner loop then fine-tunes or adapts those parameters to a specific new task, ideally in very few gradient steps or forward passes. This structure gives rise to three broad families of approaches: optimization-based methods (such as MAML, which learns an initialization that can be quickly fine-tuned), metric-based methods (such as Prototypical Networks, which learn an embedding space where task examples cluster meaningfully), and model-based methods (which use recurrent or memory-augmented architectures to encode task context directly into model state).
Meta-learning gained significant traction in the machine learning community during the 2010s, driven by advances in deep learning and growing interest in few-shot learning benchmarks. Landmark papers like Model-Agnostic Meta-Learning (MAML) in 2017 demonstrated that gradient-based meta-learning could be applied broadly across vision, reinforcement learning, and robotics tasks, sparking a wave of follow-on research.
The practical importance of meta-learning extends well beyond academic benchmarks. In real-world applications—personalized medicine, robotics, natural language processing for low-resource languages—collecting large task-specific datasets is often infeasible. Meta-learning offers a principled framework for building systems that generalize their learning ability itself, moving AI closer to the flexible, sample-efficient adaptation that characterizes human cognition. As foundation models grow larger, meta-learning principles also inform how these models are fine-tuned and prompted for new tasks.