In recent years, the discussion about systematicity of neural network learning has gained renewed interest, in particular the formal analysis of neural network behaviour. In this paper, we investigate the capability of single-cell ReLU RNN models to demonstrate precise counting behaviour. Formally, we start by characterising the semi-Dyck-1 language and semi-Dyck-1 counter machine that can be implemented by a single Rectified Linear Unit (ReLU) cell. We define three Counter Indicator Conditions (CICs) on the weights of a ReLU cell and show that fulfilling these conditions is equivalent to accepting the semi-Dyck-1 language, i.e. to perform exact counting. Empirically, we study the ability of single-cell ReLU RNNs to learn to count by traini...
Many first-order probabilistic models can be repre-sented much more compactly using aggregation oper...
In this paper, we investigate the memory properties of two popular gated units: long short term memo...
What does it mean for a neural network to become a “cardinal principal knower”? We trained a multila...
In recent years, the discussion about systematicity of neural network learning has gained renewed in...
Recently researchers have derived formal complexity analysis of analog computation in the setting of...
The broad context of this study is the investigation of the nature of computation in recurrent netwo...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
The increasing computational power and the availability of data have made it possible to train ever-...
In theory, the choice of ReLU(0) in [0, 1] for a neural network has a negligible influence both on b...
International audienceIn theory, the choice of ReLU(0) in [0, 1] for a neural network has a negligib...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
We introduce exact macroscopic on-line learning dynamics of two-layer neural networks with ReLU unit...
Artificial neural networks (ANNs) with recurrence and self-attention have been shown to be Turing-co...
Many first-order probabilistic models can be repre-sented much more compactly using aggregation oper...
In this paper, we investigate the memory properties of two popular gated units: long short term memo...
What does it mean for a neural network to become a “cardinal principal knower”? We trained a multila...
In recent years, the discussion about systematicity of neural network learning has gained renewed in...
Recently researchers have derived formal complexity analysis of analog computation in the setting of...
The broad context of this study is the investigation of the nature of computation in recurrent netwo...
© 2019 Neural information processing systems foundation. All rights reserved. We study finite sample...
The increasing computational power and the availability of data have made it possible to train ever-...
In theory, the choice of ReLU(0) in [0, 1] for a neural network has a negligible influence both on b...
International audienceIn theory, the choice of ReLU(0) in [0, 1] for a neural network has a negligib...
Recurrent Neural Networks (RNNs) are theoretically Turing-complete and established themselves as a d...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
We introduce exact macroscopic on-line learning dynamics of two-layer neural networks with ReLU unit...
Artificial neural networks (ANNs) with recurrence and self-attention have been shown to be Turing-co...
Many first-order probabilistic models can be repre-sented much more compactly using aggregation oper...
In this paper, we investigate the memory properties of two popular gated units: long short term memo...
What does it mean for a neural network to become a “cardinal principal knower”? We trained a multila...