Nowadays neural networks are a powerful tool, even if there are few mathematical results that explain the effectiveness of this approach.Until a few years ago, one of the powerful results guaranteed that any continuous function can be well approximated by a two-layers neural network with convex activation functions and enough hidden nodes.However this tells us nothing about the practical choice of the parameters.Typically the Stochastic Gradient Descent (SGD), or one of its variants, is used to update them.In the last years several results have been discovered in order to analyse the convergence of parameters using the SGD, in particular using the mean field approach.The key idea is to consider a risk function defined over a set of distribu...
Recently, neural networks (NN) with an infinite number of layers have been introduced. Especially f...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
We study the probabilistic generative models parameterized by feedforward neural networks. An attrac...
8 pages + appendix, 4 figuresInternational audienceWe analyze in a closed form the learning dynamics...
We analyze in a closed form the learning dynamics of the stochastic gradient descent (SGD) for a sin...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
In this thesis, we theoretically analyze the ability of neural networks trained by gradient descent ...
The deep learning optimization community has observed how the neural networks generalization ability...
We rigorously prove a central limit theorem for neural network models with a single hidden layer. Th...
The stochastic heavy ball method (SHB), also known as stochastic gradient descent (SGD) with Polyak'...
We present a probabilistic analysis of the long-time behaviour of the nonlocal, diffusive equations ...
Normalized gradient descent has shown substantial success in speeding up the convergence of exponen...
International audienceTraining over-parameterized neural networks involves the empirical minimizatio...
Despite the non-convex optimization landscape, over-parametrized shallow networks are able to achiev...
We study the overparametrization bounds required for the global convergence of stochastic gradient d...
Recently, neural networks (NN) with an infinite number of layers have been introduced. Especially f...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
We study the probabilistic generative models parameterized by feedforward neural networks. An attrac...
8 pages + appendix, 4 figuresInternational audienceWe analyze in a closed form the learning dynamics...
We analyze in a closed form the learning dynamics of the stochastic gradient descent (SGD) for a sin...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
In this thesis, we theoretically analyze the ability of neural networks trained by gradient descent ...
The deep learning optimization community has observed how the neural networks generalization ability...
We rigorously prove a central limit theorem for neural network models with a single hidden layer. Th...
The stochastic heavy ball method (SHB), also known as stochastic gradient descent (SGD) with Polyak'...
We present a probabilistic analysis of the long-time behaviour of the nonlocal, diffusive equations ...
Normalized gradient descent has shown substantial success in speeding up the convergence of exponen...
International audienceTraining over-parameterized neural networks involves the empirical minimizatio...
Despite the non-convex optimization landscape, over-parametrized shallow networks are able to achiev...
We study the overparametrization bounds required for the global convergence of stochastic gradient d...
Recently, neural networks (NN) with an infinite number of layers have been introduced. Especially f...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
We study the probabilistic generative models parameterized by feedforward neural networks. An attrac...