This paper presents a margin-based multiclass generalization bound for neural networks that scales with their margin-normalized "spectral complexity": their Lipschitz constant, meaning the product of the spectral norms of the weight matrices, times a certain correction factor. This bound is empirically investigated for a standard AlexNet network trained with SGD on the MNIST and CIFAR10 datasets, with both original and random labels; the bound, the Lipschitz constants, and the excess risks are all in direct correlation, suggesting both that SGD selects predictors whose complexity scales with the difficulty of the learning task, and secondly that the presented bound is sensitive to this complexity
This paper tackles the problem of Lipschitz regularization of Convolutional Neural Networks. Lipschi...
Effective regularisation of neural networks is essential to combat overfitting due to the large numb...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
This paper presents a margin-based multiclass generalization bound for neural networks that scales w...
Exciting new work on the generalization bounds for neural networks (NN) givenby Neyshabur et al. ,...
Weight norm $\|w\|$ and margin $\gamma$ participate in learning theory via the normalized margin $\g...
We investigate the effect of explicitly enforcing the Lipschitz continuity of neural networks with r...
We study the sample complexity of learning neural networks by providing new bounds on their Rademach...
Generalization bounds depending on the margin of a classifier are a relatively recent development. T...
Anumber of results have bounded generalization of a classi er in terms of its margin on the training...
We show generalisation error bounds for deep learning with two main improvements over the state of t...
Normalized gradient descent has shown substantial success in speeding up the convergence of exponen...
We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding ...
International audienceThe stability of neural networks with respect to adversarial perturbations has...
Abstract. A class of neuralmodels isintroduced inwhichthe topology of the neural network has been ge...
This paper tackles the problem of Lipschitz regularization of Convolutional Neural Networks. Lipschi...
Effective regularisation of neural networks is essential to combat overfitting due to the large numb...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
This paper presents a margin-based multiclass generalization bound for neural networks that scales w...
Exciting new work on the generalization bounds for neural networks (NN) givenby Neyshabur et al. ,...
Weight norm $\|w\|$ and margin $\gamma$ participate in learning theory via the normalized margin $\g...
We investigate the effect of explicitly enforcing the Lipschitz continuity of neural networks with r...
We study the sample complexity of learning neural networks by providing new bounds on their Rademach...
Generalization bounds depending on the margin of a classifier are a relatively recent development. T...
Anumber of results have bounded generalization of a classi er in terms of its margin on the training...
We show generalisation error bounds for deep learning with two main improvements over the state of t...
Normalized gradient descent has shown substantial success in speeding up the convergence of exponen...
We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding ...
International audienceThe stability of neural networks with respect to adversarial perturbations has...
Abstract. A class of neuralmodels isintroduced inwhichthe topology of the neural network has been ge...
This paper tackles the problem of Lipschitz regularization of Convolutional Neural Networks. Lipschi...
Effective regularisation of neural networks is essential to combat overfitting due to the large numb...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...