This article studies the infinite-width limit of deep feedforward neural networks whose weights are dependent, and modelled via a mixture of Gaussian distributions. Each hidden node of the network is assigned a nonnegative random variable that controls the variance of the outgoing weights of that node. We make minimal assumptions on these per-no de random variables: they are iid and their sum, in each layer, converges to some finite random variable in the infinite-width limit. Under this model, we show that each layer of the infinite-width neural network can be characterised by two simple quantities: a non-negative scalar parameter and a Levy measure on the positive reals. If the scalar parameters are strictly positive and the Levy measures...
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound...
A recent series of theoretical works showed that the dynamics of neural networks with a certain init...
Deep neural networks (DNNs) are powerful tools for compressing and distilling information. Their sca...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
It took until the last decade to finally see a machine match human performance on essentially any ta...
Choosing appropriate architectures and regular-ization strategies for deep networks is crucial to go...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
A general analysis of the limiting distribution of neural network functions is performed, with empha...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
This thesis aims to study recent theoretical work in machine learning research that seeks to better ...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
Many modern machine learning methods, including deep neural networks, utilize a discrete sequence of...
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound...
A recent series of theoretical works showed that the dynamics of neural networks with a certain init...
Deep neural networks (DNNs) are powerful tools for compressing and distilling information. Their sca...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
It took until the last decade to finally see a machine match human performance on essentially any ta...
Choosing appropriate architectures and regular-ization strategies for deep networks is crucial to go...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
We show that the output of a (residual) convolutional neural network (CNN) with an appropriate prior...
A general analysis of the limiting distribution of neural network functions is performed, with empha...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
This thesis aims to study recent theoretical work in machine learning research that seeks to better ...
International audienceThe connection between Bayesian neural networks and Gaussian processes gained ...
Many modern machine learning methods, including deep neural networks, utilize a discrete sequence of...
Given any deep fully connected neural network, initialized with random Gaussian parameters, we bound...
A recent series of theoretical works showed that the dynamics of neural networks with a certain init...
Deep neural networks (DNNs) are powerful tools for compressing and distilling information. Their sca...