International audienceGiven a training set, a loss function, and a neural network architecture, it is often taken for granted that optimal network parameters exist, and a common practice is to apply available optimization algorithms to search for them. In this work, we show that the existence of an optimal solution is not always guaranteed, especially in the context of {\em sparse} ReLU neural networks. In particular, we first show that optimization problems involving deep networks with certain sparsity patterns do not always have optimal parameters, and that optimization algorithms may then diverge. Via a new topological relation between sparse ReLU neural networks and their linear counterparts, we derive --using existing tools from real ...
The motivation for this work is to improve the performance of deep neural networks through the optim...
Understanding the computational complexity of training simple neural networks with rectified linear ...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
In the interest of reproducible research, this is exactly the version of the code used to generate t...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
The over-parameterization of neural networks and the local optimality of backpropagation algorithm h...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
Abstract Sparse neural networks can achieve performance comparable to fully connected networks but n...
Sparse neural networks are effective approaches to reduce the resource requirements for the deployme...
Convex $\ell_1$ regularization using an infinite dictionary of neurons has been suggested for constr...
The leaky ReLU network with a group sparse regularization term has been widely used in the recent ye...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
This work investigates Sparse Neural Networks, which are artificial neural information processing sy...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
Recently, sparse training methods have started to be established as a de facto approach for training...
The motivation for this work is to improve the performance of deep neural networks through the optim...
Understanding the computational complexity of training simple neural networks with rectified linear ...
We contribute to a better understanding of the class of functions that is represented by a neural ne...
In the interest of reproducible research, this is exactly the version of the code used to generate t...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
The over-parameterization of neural networks and the local optimality of backpropagation algorithm h...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
Abstract Sparse neural networks can achieve performance comparable to fully connected networks but n...
Sparse neural networks are effective approaches to reduce the resource requirements for the deployme...
Convex $\ell_1$ regularization using an infinite dictionary of neurons has been suggested for constr...
The leaky ReLU network with a group sparse regularization term has been widely used in the recent ye...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
This work investigates Sparse Neural Networks, which are artificial neural information processing sy...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
Recently, sparse training methods have started to be established as a de facto approach for training...
The motivation for this work is to improve the performance of deep neural networks through the optim...
Understanding the computational complexity of training simple neural networks with rectified linear ...
We contribute to a better understanding of the class of functions that is represented by a neural ne...