We rigorously prove a central limit theorem for neural network models with a single hidden layer. The central limit theorem is proven in the asymptotic regime of simultaneously (A) large numbers of hidden units and (B) large numbers of stochastic gradient descent training iterations. Our result describes the neural network’s fluctuations around its mean-field limit. The fluctuations have a Gaussian distribution and satisfy a stochastic partial differential equation. The proof relies upon weak convergence methods from stochastic analysis. In particular, we prove relative compactness for the sequence of processes and uniqueness of the limiting process in a suitable Sobolev space
We consider a system of N neurons, each spiking randomly with rate depending on its membrane potenti...
: We prove a central limit theorem for the finite dimensional marginals of the Gibbs distribution of...
42 pages, 4 figuresThis paper establishes limit theorems for a class of stochastic hybrid systems (c...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
In this work, we consider a wide two-layer neural network and study the behavior of its empirical we...
In a previous paper, it has been shown that the mean-field limit of spatially extended Hawkes proces...
permits unrestricted use, distribution, and reproduction in any medium, provided the original work i...
Nowadays neural networks are a powerful tool, even if there are few mathematical results that explai...
We prove a Quantitative Functional Central Limit Theorem for one-hidden-layer neural networks with g...
We review a recent approach to the mean-field limits in neural networks that takes into account the ...
We study the overparametrization bounds required for the global convergence of stochastic gradient d...
Mathematical models of biological neural networks are associated to a rich and complex class of stoc...
The stochastic heavy ball method (SHB), also known as stochastic gradient descent (SGD) with Polyak'...
In this work we study of the dynamics of large-size random neural networks. Different methods have b...
We consider an autoregressive process with a nonlinear regression function that is modeled by a feed...
We consider a system of N neurons, each spiking randomly with rate depending on its membrane potenti...
: We prove a central limit theorem for the finite dimensional marginals of the Gibbs distribution of...
42 pages, 4 figuresThis paper establishes limit theorems for a class of stochastic hybrid systems (c...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
In this work, we consider a wide two-layer neural network and study the behavior of its empirical we...
In a previous paper, it has been shown that the mean-field limit of spatially extended Hawkes proces...
permits unrestricted use, distribution, and reproduction in any medium, provided the original work i...
Nowadays neural networks are a powerful tool, even if there are few mathematical results that explai...
We prove a Quantitative Functional Central Limit Theorem for one-hidden-layer neural networks with g...
We review a recent approach to the mean-field limits in neural networks that takes into account the ...
We study the overparametrization bounds required for the global convergence of stochastic gradient d...
Mathematical models of biological neural networks are associated to a rich and complex class of stoc...
The stochastic heavy ball method (SHB), also known as stochastic gradient descent (SGD) with Polyak'...
In this work we study of the dynamics of large-size random neural networks. Different methods have b...
We consider an autoregressive process with a nonlinear regression function that is modeled by a feed...
We consider a system of N neurons, each spiking randomly with rate depending on its membrane potenti...
: We prove a central limit theorem for the finite dimensional marginals of the Gibbs distribution of...
42 pages, 4 figuresThis paper establishes limit theorems for a class of stochastic hybrid systems (c...