Convex $\ell_1$ regularization using an infinite dictionary of neurons has been suggested for constructing neural networks with desired approximation guarantees, but can be affected by an arbitrary amount of over-parametrization. This can lead to a loss of sparsity and result in networks with too many active neurons for the given data, in particular if the number of data samples is large. As a remedy, in this paper, a nonconvex regularization method is investigated in the context of shallow ReLU networks: We prove that in contrast to the convex approach, any resulting (locally optimal) network is finite even in the presence of infinite data (i.e., if the data distribution is known and the limiting case of infinite samples is considered). Mo...
We consider a deep neural network estimator based on empirical risk minimization with l_1-regulariza...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
International audienceSparsifying deep neural networks is of paramount interest in many areas, espec...
This work investigates Sparse Neural Networks, which are artificial neural information processing sy...
International audienceWe consider neural networks with a single hidden layer and non-decreasing homo...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
The over-parameterization of neural networks and the local optimality of backpropagation algorithm h...
International audienceGiven a training set, a loss function, and a neural network architecture, it i...
We provide novel guaranteed approaches for training feedforward neural networks with sparse connecti...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Recently, sparse training methods have started to be established as a de facto approach for training...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
Two aspects of neural networks that have been extensively studied in the recent literature are their...
Recent theoretical works on over-parameterized neural nets have focused on two aspects: optimization...
We consider a deep neural network estimator based on empirical risk minimization with l_1-regulariza...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
International audienceSparsifying deep neural networks is of paramount interest in many areas, espec...
This work investigates Sparse Neural Networks, which are artificial neural information processing sy...
International audienceWe consider neural networks with a single hidden layer and non-decreasing homo...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
The over-parameterization of neural networks and the local optimality of backpropagation algorithm h...
International audienceGiven a training set, a loss function, and a neural network architecture, it i...
We provide novel guaranteed approaches for training feedforward neural networks with sparse connecti...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
In this paper, we address the challenging task of simultaneously optimizing (i) the weights of a neu...
Recently, sparse training methods have started to be established as a de facto approach for training...
The practice of deep learning has shown that neural networks generalize remarkably well even with an...
Two aspects of neural networks that have been extensively studied in the recent literature are their...
Recent theoretical works on over-parameterized neural nets have focused on two aspects: optimization...
We consider a deep neural network estimator based on empirical risk minimization with l_1-regulariza...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
International audienceSparsifying deep neural networks is of paramount interest in many areas, espec...