Inspired by the concept of preconditioning, we propose a novel method to increase adaptation speed for gradient-based meta-learning methods without incurring extra parameters. We demonstrate that recasting the optimization problem to a non-linear least-squares formulation provides a principled way to actively enforce a $\textit{well-conditioned}$ parameter space for meta-learning models based on the concepts of the condition number and local curvature. Our comprehensive evaluations show that the proposed method significantly outperforms its unconstrained counterpart especially during initial adaptation steps, while achieving comparable or better overall results on several few-shot classification tasks -- creating the possibility of dynamica...
We aim to design adaptive online learning algorithms that take advantage of any special structure t...
Finding neural network weights that generalize well from small datasets is difficult. A promising ap...
Meta-learning algorithms leverage regularities that are present on a set of tasks to speed up and im...
Gradient-based meta-learning techniques aim to distill useful prior knowledge from a set of training...
In this paper, we consider the framework of multi-task representation (MTR) learning where the goal ...
Intelligent agents should have the ability to leverage knowledge from previously learned tasks in or...
While developments in machine learning led to impressive performance gains on big data, many human s...
International audienceIn this paper, we review the recent advances in meta-learning theory and show ...
In this work we provide an analysis of the distribution of the post-adaptation parameters of Gradien...
The prototypical network is a prototype classifier based on meta-learning and is widely used for few...
The aim of Few-Shot learning methods is to train models which can easily adapt to previously unseen ...
Model-agnostic meta-learning (MAML) is a meta-learning technique to train a model on a multitude of ...
Humans have a remarkable ability to learn new concepts from only a few examples and quickly adapt to...
Meta-Learning, or so-called Learning to learn, has become another important research branch in Machi...
Black-Box Tuning (BBT) is a derivative-free approach to optimize continuous prompt tokens prepended ...
We aim to design adaptive online learning algorithms that take advantage of any special structure t...
Finding neural network weights that generalize well from small datasets is difficult. A promising ap...
Meta-learning algorithms leverage regularities that are present on a set of tasks to speed up and im...
Gradient-based meta-learning techniques aim to distill useful prior knowledge from a set of training...
In this paper, we consider the framework of multi-task representation (MTR) learning where the goal ...
Intelligent agents should have the ability to leverage knowledge from previously learned tasks in or...
While developments in machine learning led to impressive performance gains on big data, many human s...
International audienceIn this paper, we review the recent advances in meta-learning theory and show ...
In this work we provide an analysis of the distribution of the post-adaptation parameters of Gradien...
The prototypical network is a prototype classifier based on meta-learning and is widely used for few...
The aim of Few-Shot learning methods is to train models which can easily adapt to previously unseen ...
Model-agnostic meta-learning (MAML) is a meta-learning technique to train a model on a multitude of ...
Humans have a remarkable ability to learn new concepts from only a few examples and quickly adapt to...
Meta-Learning, or so-called Learning to learn, has become another important research branch in Machi...
Black-Box Tuning (BBT) is a derivative-free approach to optimize continuous prompt tokens prepended ...
We aim to design adaptive online learning algorithms that take advantage of any special structure t...
Finding neural network weights that generalize well from small datasets is difficult. A promising ap...
Meta-learning algorithms leverage regularities that are present on a set of tasks to speed up and im...