Several regularization methods have recently been introduced which force the latent activations of an autoencoder or deep neural network to conform to either a Gaussian or hyperspherical distribution, or to minimize the implicit rank of the distribution in latent space. In the present work, we introduce a novel regularizing loss function which simulates a pairwise repulsive force between items and an attractive force of each item toward the origin. We show that minimizing this loss function in isolation achieves a hyperspherical distribution. Moreover, when used as a regularizing term, the scaling factor can be adjusted to allow greater flexibility and tolerance of eccentricity, thus allowing the latent variables to be stratified according ...
Recent advances in deep learning have relied on large, labelled datasets to train high-capacity mode...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
Over the last decade, learning theory performed significant progress in the development of sophistic...
Injecting noise within gradient descent has several desirable features. In this paper, we explore no...
Neural networks are more expressive when they have multiple layers. In turn, conventional training m...
Is it possible to train several classifiers to perform meaningful crowd-sourcing to produce a better...
This work used the Cirrus UK National Tier-2 HPC Service at EPCC (http://www.cirrus.ac.uk). Access g...
We show that training common regularized autoencoders resembles clustering, because it amounts to fi...
In Part I of the thesis, we present a body of work analyzing and deriving data-centric regularizatio...
We study the loss surface of DNNs with $L_{2}$ regularization. We show that the loss in terms of the...
We had previously shown that regularization principles lead to ap-proximation schemes that are equiv...
Deep learning based techniques achieve state-of-the-art results in a wide range of image reconstruct...
The volume of the distribution of weight sets associated with a loss value may be the source of impl...
In this study, the problem of computing a sparse representation of multi-dimensional visual data is ...
We introduce HyperMorph, a framework that facilitates efficient hyperparameter tuning in learning-ba...
Recent advances in deep learning have relied on large, labelled datasets to train high-capacity mode...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
Over the last decade, learning theory performed significant progress in the development of sophistic...
Injecting noise within gradient descent has several desirable features. In this paper, we explore no...
Neural networks are more expressive when they have multiple layers. In turn, conventional training m...
Is it possible to train several classifiers to perform meaningful crowd-sourcing to produce a better...
This work used the Cirrus UK National Tier-2 HPC Service at EPCC (http://www.cirrus.ac.uk). Access g...
We show that training common regularized autoencoders resembles clustering, because it amounts to fi...
In Part I of the thesis, we present a body of work analyzing and deriving data-centric regularizatio...
We study the loss surface of DNNs with $L_{2}$ regularization. We show that the loss in terms of the...
We had previously shown that regularization principles lead to ap-proximation schemes that are equiv...
Deep learning based techniques achieve state-of-the-art results in a wide range of image reconstruct...
The volume of the distribution of weight sets associated with a loss value may be the source of impl...
In this study, the problem of computing a sparse representation of multi-dimensional visual data is ...
We introduce HyperMorph, a framework that facilitates efficient hyperparameter tuning in learning-ba...
Recent advances in deep learning have relied on large, labelled datasets to train high-capacity mode...
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
Over the last decade, learning theory performed significant progress in the development of sophistic...