Overwhelming theoretical and empirical evidence shows that mildly overparametrized neural networks -- those with more connections than the size of the training data -- are often able to memorize the training data with $100\%$ accuracy. This was rigorously proved for networks with sigmoid activation functions and, very recently, for ReLU activations. Addressing a 1988 open question of Baum, we prove that this phenomenon holds for general multilayered perceptrons, i.e. neural networks with threshold activation functions, or with any mix of threshold and ReLU activations. Our construction is probabilistic and exploits sparsity
Recurrent neural networks have been shown to be able to store memory patterns as fixed point attract...
We present a model of long term memory : learning within irreversible bounds. The best bound values ...
This article discusses a number of reasons why the use of non-monotonic functions as activation func...
In this article we present new results on neural networks with linear threshold activation functions...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
It is well-known that neural networks are computationally hard to train. On the other hand, in pract...
Threshold-linear (graded response) units approximate the real firing behaviour of pyramidal neurons ...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
Determining the memory capacity of two layer neural networks with $m$ hidden neurons and input dimen...
This paper deals with learnability of concept classes defined by neural networks, showing the hardne...
MEng (Computer and Electronic Engineering), North-West University, Potchefstroom CampusThe ability o...
International audienceArtificial Neural Networks (ANNs) trained with Backpropagation and Stochastic ...
This paper shows that neural networks which use continuous activation functions have VC dimension at...
The task of a neural associative memory is to retrieve a set of previously memorized patterns from t...
We deal with computational issues of loading a fixed-architecture neural network with a set of posit...
Recurrent neural networks have been shown to be able to store memory patterns as fixed point attract...
We present a model of long term memory : learning within irreversible bounds. The best bound values ...
This article discusses a number of reasons why the use of non-monotonic functions as activation func...
In this article we present new results on neural networks with linear threshold activation functions...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
It is well-known that neural networks are computationally hard to train. On the other hand, in pract...
Threshold-linear (graded response) units approximate the real firing behaviour of pyramidal neurons ...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
Determining the memory capacity of two layer neural networks with $m$ hidden neurons and input dimen...
This paper deals with learnability of concept classes defined by neural networks, showing the hardne...
MEng (Computer and Electronic Engineering), North-West University, Potchefstroom CampusThe ability o...
International audienceArtificial Neural Networks (ANNs) trained with Backpropagation and Stochastic ...
This paper shows that neural networks which use continuous activation functions have VC dimension at...
The task of a neural associative memory is to retrieve a set of previously memorized patterns from t...
We deal with computational issues of loading a fixed-architecture neural network with a set of posit...
Recurrent neural networks have been shown to be able to store memory patterns as fixed point attract...
We present a model of long term memory : learning within irreversible bounds. The best bound values ...
This article discusses a number of reasons why the use of non-monotonic functions as activation func...