We address the following question: How redundant is the parameterisation of ReLU networks? Specifically, we consider transformations of the weight space which leave the function implemented by the network intact. Two such transformations are known for feed-forward architectures: permutation of neurons within a layer, and positive scaling of all incoming weights of a neuron coupled with inverse scaling of its outgoing weights. In this work, we show for architectures with non-increasing widths that permutation and scaling are in fact the only function-preserving weight transformations. For any eligible architecture we give an explicit construction of a neural network such that any other network that implements the same function can be obt...
We study the loss surface of neural networks equipped with a hinge loss criterion and ReLU or leaky ...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
Understanding the computational complexity of training simple neural networks with rectified linear ...
We address the following question: How redundant is the parameterisation of ReLU networks? Specific...
Neural networks with the Rectified Linear Unit (ReLU) nonlinearity are described by a vector of para...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
The possibility for one to recover the parameters-weights and biases-of a neural network thanks to t...
It is well-known that the parameterized family of functions representable by fully-connected feedfor...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
We contribute to a better understanding of the class of functions that can be represented by a neura...
This paper explores the expressive power of deep neural networks through the framework of function c...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
We deal with two complementary questions about approximation properties of ReLU networks. First, we ...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
We study the loss surface of neural networks equipped with a hinge loss criterion and ReLU or leaky ...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
Understanding the computational complexity of training simple neural networks with rectified linear ...
We address the following question: How redundant is the parameterisation of ReLU networks? Specific...
Neural networks with the Rectified Linear Unit (ReLU) nonlinearity are described by a vector of para...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
The possibility for one to recover the parameters-weights and biases-of a neural network thanks to t...
It is well-known that the parameterized family of functions representable by fully-connected feedfor...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
We contribute to a better understanding of the class of functions that can be represented by a neura...
This paper explores the expressive power of deep neural networks through the framework of function c...
We solve an open question from Lu et al. (2017), by showing that any target network with inputs in $...
We deal with two complementary questions about approximation properties of ReLU networks. First, we ...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
We study the loss surface of neural networks equipped with a hinge loss criterion and ReLU or leaky ...
By applying concepts from the statistical physics of learning, we study layered neural networks of r...
Understanding the computational complexity of training simple neural networks with rectified linear ...