As its name implies, this optimization-based meta learning algorithm is model-agnostic. This makes it compatible with any model trained using gradient descent and suitable for solving various learning problems, such as classification, regression and reinforcement learning.8
The core idea behind MAML is to train the model’s initial parameters in a way that a few gradient updates will result in rapid learning on a new task. The goal is to determine model parameters that are sensitive to changes in a task such that minor changes to those parameters lead to major improvements in the task’s loss function. Meta-optimization across tasks is done using stochastic gradient descent (SGD).8
Unlike gradient descent, which computes derivatives to optimize a model’s parameters for a certain task, MAML computes second derivatives to optimize a model’s initial parameters for task-specific optimization. A modified version of model-agnostic meta learning, known as first-order MAML or FOMAML, omits second derivatives for a less computationally expensive process.8