The weight matrix (WM) of a neural network (NN) is its program. The programs of many traditional NNs are learned through gradient descent in some error function, then remain fixed. The WM of a self-referential NN, however, can keep rapidly modifying all of itself during runtime. In principle, such NNs can meta-learn to learn, and meta-meta-learn to meta-learn to learn, and so on, in the sense of recursive self-improvement. While NN architectures potentially capable of implementing such behaviour have been proposed since the '90s, there have been few if any practical studies. Here we revisit such NNs, building upon recent successes of fast weight programmers and closely related linear Transformers. We propose a scalable self-referential WM (...
In this paper we introduce a novel neural network architecture, in which weight matrices are re-para...
Compositionality is a basic structural feature of both biological and artificial neural networks. Le...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
Weight modifications in traditional neural nets are computed by hard-wired algorithms. Without excep...
Traditional artificial neural networks cannot reflect about their own weight modification algorithm....
The transformer architecture and variants presented remarkable success across many machine learning ...
We propose self-adaptive training -- a unified training algorithm that dynamically calibrates and en...
Deep Reinforcement Learning has demonstrated the potential of neural networks tuned with gradient de...
Schiller UD, Steil JJ. On the weight dynamcis of recurrent learning. In: Verleysen M, ed. Proc. Euro...
Self-Supervised Learning (SSL) has been shown to learn useful and information-preserving representat...
Recently the surprising discovery of the Bootstrap Your Own Latent (BYOL) method by Grill et al. sho...
The State of the Art of the young domain of Meta-Learning [3] is held by the connectionist approach....
The brain processes information through many layers of neurons. This deep architecture is representa...
Rumelhart, Hinton and Williams [Rumelhart et al. 86] describe a learning procedure for layered netwo...
We have recently investigated a new way of conceptualizing the inferential capacities of non-linear ...
In this paper we introduce a novel neural network architecture, in which weight matrices are re-para...
Compositionality is a basic structural feature of both biological and artificial neural networks. Le...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...
Weight modifications in traditional neural nets are computed by hard-wired algorithms. Without excep...
Traditional artificial neural networks cannot reflect about their own weight modification algorithm....
The transformer architecture and variants presented remarkable success across many machine learning ...
We propose self-adaptive training -- a unified training algorithm that dynamically calibrates and en...
Deep Reinforcement Learning has demonstrated the potential of neural networks tuned with gradient de...
Schiller UD, Steil JJ. On the weight dynamcis of recurrent learning. In: Verleysen M, ed. Proc. Euro...
Self-Supervised Learning (SSL) has been shown to learn useful and information-preserving representat...
Recently the surprising discovery of the Bootstrap Your Own Latent (BYOL) method by Grill et al. sho...
The State of the Art of the young domain of Meta-Learning [3] is held by the connectionist approach....
The brain processes information through many layers of neurons. This deep architecture is representa...
Rumelhart, Hinton and Williams [Rumelhart et al. 86] describe a learning procedure for layered netwo...
We have recently investigated a new way of conceptualizing the inferential capacities of non-linear ...
In this paper we introduce a novel neural network architecture, in which weight matrices are re-para...
Compositionality is a basic structural feature of both biological and artificial neural networks. Le...
Abstract We present weight normalization: a reparameterization of the weight vectors in a neural net...