Recent developments in few-shot learning have shown that during fast adaption, gradient-based meta-learners mostly rely on embedding features of powerful pretrained networks. This leads us to research ways to effectively adapt features and utilize the meta-learner's full potential. Here, we demonstrate the effectiveness of hypernetworks in this context. We propose a soft row-sharing hypernetwork architecture and show that training the hypernetwork with a variant of MAML is tightly linked to meta-learning a curvature matrix used to condition gradients during fast adaptation. We achieve similar results as state-of-art model-agnostic methods in the overparametrized case, while outperforming many MAML variants without using different optimizati...
Deep neural networks can achieve great successes when presented with large data sets and sufficient ...
Few-shot learning for neural networks (NNs) is an important problem that aims to train NNs with a fe...
In order to make predictions with high accuracy, conventional deep learning systems require large tr...
Recent developments in few-shot learning have shown that during fast adaption, gradient-based meta-l...
The aim of Few-Shot learning methods is to train models which can easily adapt to previously unseen ...
Despite huge progress in artificial intelligence, the ability to quickly learn from few examples is ...
Machine learning algorithms and systems are progressively becoming part of our societies, leading to...
Humans have a remarkable ability to learn new concepts from only a few examples and quickly adapt to...
MasterDeep learning has been tremendously successful in many difficult tasks including image classi...
The State of the Art of the young domain of Meta-Learning [3] is held by the connectionist approach....
© 2019, Springer Nature Switzerland AG. In the last few years, we have witnessed a resurgence of int...
Model-agnostic meta-learning (MAML) is a meta-learning technique to train a model on a multitude of ...
Finding neural network weights that generalize well from small datasets is difficult. A promising ap...
Few-Shot learning (FSL) jest problemem zorientowanym na uczenie się z ograniczonej ilości danych. Ce...
We introduce a framework based on bilevel programming that unifies gradient-based hyperparameter opt...
Deep neural networks can achieve great successes when presented with large data sets and sufficient ...
Few-shot learning for neural networks (NNs) is an important problem that aims to train NNs with a fe...
In order to make predictions with high accuracy, conventional deep learning systems require large tr...
Recent developments in few-shot learning have shown that during fast adaption, gradient-based meta-l...
The aim of Few-Shot learning methods is to train models which can easily adapt to previously unseen ...
Despite huge progress in artificial intelligence, the ability to quickly learn from few examples is ...
Machine learning algorithms and systems are progressively becoming part of our societies, leading to...
Humans have a remarkable ability to learn new concepts from only a few examples and quickly adapt to...
MasterDeep learning has been tremendously successful in many difficult tasks including image classi...
The State of the Art of the young domain of Meta-Learning [3] is held by the connectionist approach....
© 2019, Springer Nature Switzerland AG. In the last few years, we have witnessed a resurgence of int...
Model-agnostic meta-learning (MAML) is a meta-learning technique to train a model on a multitude of ...
Finding neural network weights that generalize well from small datasets is difficult. A promising ap...
Few-Shot learning (FSL) jest problemem zorientowanym na uczenie się z ograniczonej ilości danych. Ce...
We introduce a framework based on bilevel programming that unifies gradient-based hyperparameter opt...
Deep neural networks can achieve great successes when presented with large data sets and sufficient ...
Few-shot learning for neural networks (NNs) is an important problem that aims to train NNs with a fe...
In order to make predictions with high accuracy, conventional deep learning systems require large tr...