It took until the last decade to finally see a machine match human performance on essentially any task related to vision or natural language understanding. Most of these successes were achieved by neural networks (NNs), a class of algorithms for finding patterns in large swaths of data. The progress since 2020 in particular has been driven by: (i) growing the number of parameters NNs can use to make predictions, and (ii) increasing the amount of data used to optimise these parameters. The race for scale has been fuelled by the discovery of scaling laws, an empirical phenomenon where the number of errors a model makes decays as a power law of the dataset size and NN parameter count. This thesis is devoted to understanding how parameter cou...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
The successful application of ConvNets and other neural architectures to computer vision is central ...
Two distinct limits for deep learning have been derived as the network width h -> infinity, dependin...
Deep neural networks have had tremendous success in a wide range of applications where they achieve ...
A remarkable characteristic of overparameterized deep neural networks (DNNs) is that their accuracy ...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
One of the most important aspects of any machine learning paradigm is how it scales according to pro...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
We study the compute-optimal trade-off between model and training data set sizes for large neural ne...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
International audienceDeep neural networks of sizes commonly encountered in practice are proven to c...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
For many reasons, neural networks have become very popular AI machine learning models. Two of the mo...
Deep ResNets are recognized for achieving state-of-the-art results in complex machine learning tasks...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
The successful application of ConvNets and other neural architectures to computer vision is central ...
Two distinct limits for deep learning have been derived as the network width h -> infinity, dependin...
Deep neural networks have had tremendous success in a wide range of applications where they achieve ...
A remarkable characteristic of overparameterized deep neural networks (DNNs) is that their accuracy ...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Machine learning, and in particular neural network models, have revolutionized fields such as image,...
One of the most important aspects of any machine learning paradigm is how it scales according to pro...
Recent years have witnessed an increasing interest in the correspondence between infinitely wide net...
We study the compute-optimal trade-off between model and training data set sizes for large neural ne...
We investigate deep Bayesian neural networks with Gaussian priors on the weights and ReLU-like nonli...
International audienceDeep neural networks of sizes commonly encountered in practice are proven to c...
International audienceWe investigate deep Bayesian neural networks with Gaussian priors on the weigh...
For many reasons, neural networks have become very popular AI machine learning models. Two of the mo...
Deep ResNets are recognized for achieving state-of-the-art results in complex machine learning tasks...
This article studies the infinite-width limit of deep feedforward neural networks whose weights are ...
The successful application of ConvNets and other neural architectures to computer vision is central ...
Two distinct limits for deep learning have been derived as the network width h -> infinity, dependin...