Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks
This paper introduces MAML (Model-Agnostic Meta-learning Algorithm), an algorithm that works seamlessly with traditional reinforcement learning, regression, classification, and gradient descent. It aims to solve few-shot learning problems and enable quick adaptation to new tasks through "fast adaptation".
This structure reflects the distribution of overall tasks through the base-level learner. Inspired by this idea, there is a study called "Meta-Learning in Neural Networks: A Survey" that applies this concept to optimization.
Introduction
Human intelligence performs various tasks by learning from just a few examples. This algorithm for deep learning was inspired by that concept. However, MAML takes a different approach from traditional meta-learning, which focuses on function optimization, learning rates, and hyperparameters.
Meta Learning Problem Set-Up
Few-shot learning (FSL) trains models using minimal examples. MAML implements this approach through meta-learning processes that occur before the main training—known as "fast adaptation." This section covers the formulation and setup of this process.
is a task that the model should be able to solve, and it contains:
For a model, this is expressed as a function that maps observations to outputs .
is the loss function—reflecting the distribution of .
is the initial observation.
The loss value is the feedback for a specific task.
is the transition distribution associated with , .
is the episode length, representing the total number of time steps during which an model must take actions or proceed.
In this meta-learning scenario, a model must adapt to the distribution of . Similar to the k-shot framework, the model performs drawn from , where the response is derived from training samples and .
Algorithm
MAML's mechanism performs two types of learning in one episode. This is commonly explained through the concepts of inner loop with local parameters and outer loop with meta parameters.
In the inner loop, the model updates local parameters for a specific task using the local learning rate .
In the outer loop that follows, the model updates the global/meta parameters that apply across all tasks using the global learning rate .
The local and global parameters maintain identical shapes.
represents the distribution over tasks .
Initialize randomly.
Repeat until done:
Sample batch of tasks
For each , do the following:
Evaluate gradient with respect to samples and adapt parameters:
Update
The experimental section of this paper demonstrates that simultaneous execution of meta-level learning and base-level learning has a positive effect on optimization between the two learners. The analysis of these results shows that MAML can converge in fewer steps as it avoids overfitting and considers the distribution and representation between tasks.
Last updated