We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding of how these are affected by the architecture and type of norm constraint, for the simple class of scalar-valued one-hidden-layer networks, and inputs bounded in Euclidean norm. We begin by proving that in general, controlling the spectral norm of the hidden layer weight matrix is insufficient to get uniform convergence guarantees (independent of the network width), while a stronger Frobenius norm control is sufficient, extending and improving on previous work. Motivated by the proof constructions, we identify and analyze two important settings where (perhaps surprisingly) a mere spectral norm control turns out to be sufficient: First, when t...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
In many contexts, simpler models are preferable to more complex models and the control of this model...
We study the effect of normalization on the layers of deep neural networks of feed-forward type. A g...
We study the sample complexity of learning neural networks by providing new bounds on their Rademach...
We investigate the sample complexity of bounded two-layer neural networks using different activation...
Analysing Generalisation Error Bounds for Convolutional Neural Networks Abstract: Convolutional neur...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
We show generalisation error bounds for deep learning with two main improvements over the state of t...
AbstractThis paper shows that neural networks which use continuous activation functions have VC dime...
This paper presents a margin-based multiclass generalization bound for neural networks that scales w...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
Recent work by Jacot et al. (2018) has shown that training a neural network using gradient descent i...
The universal approximation theorem is generalised to uniform convergence on the (noncompact) input ...
Learning from data formalized as a minimization of a regularized empirical error is studied in terms...
We develop a mathematically rigorous framework for multilayer neural networks in the mean field regi...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
In many contexts, simpler models are preferable to more complex models and the control of this model...
We study the effect of normalization on the layers of deep neural networks of feed-forward type. A g...
We study the sample complexity of learning neural networks by providing new bounds on their Rademach...
We investigate the sample complexity of bounded two-layer neural networks using different activation...
Analysing Generalisation Error Bounds for Convolutional Neural Networks Abstract: Convolutional neur...
We consider the algorithmic problem of finding the optimal weights and biases for a two-layer fully ...
We show generalisation error bounds for deep learning with two main improvements over the state of t...
AbstractThis paper shows that neural networks which use continuous activation functions have VC dime...
This paper presents a margin-based multiclass generalization bound for neural networks that scales w...
The general features of the optimization problem for the case of overparametrized nonlinear networks...
Recent work by Jacot et al. (2018) has shown that training a neural network using gradient descent i...
The universal approximation theorem is generalised to uniform convergence on the (noncompact) input ...
Learning from data formalized as a minimization of a regularized empirical error is studied in terms...
We develop a mathematically rigorous framework for multilayer neural networks in the mean field regi...
In this note, we study how neural networks with a single hidden layer and ReLU activation interpolat...
In many contexts, simpler models are preferable to more complex models and the control of this model...
We study the effect of normalization on the layers of deep neural networks of feed-forward type. A g...