In the present work we study In some areas, artificial feed forward neural networks are still a competitive machine learning model. Unfortunately they tend to overfit the training data, which limits their ability to generalize. We study methods for regularization based on enforcing internal structure of the network. We analyze internal representations using a theoretical model based on information theory. Based on this study, we propose a regularizer that minimizes the overall entropy of internal representations. The entropy-based regularizer is computationally demanding and we use it primarily as a theoretical motivation. To develop an efficient and flexible implementation, we design a Gaussian mixture model of activations. In the experime...
Unsupervised neural networks, such as restricted Boltzmann machines (RBMs) and deep belief networks ...
The aim of this paper is to introduce two widely applicable regularization methods based on the dire...
Regularization of neural networks can alleviate overfitting in the training phase. Current regulariz...
In the present work we study In some areas, artificial feed forward neural networks are still a comp...
International audienceStudies on generalization performance of machine learning algorithms under the...
The large capacity of neural networks enables them to learn complex functions. To avoid overfitting,...
As for Gaussian mixture modeling, the key problem is to select the number of Gaussians in the mixtur...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Abstract. In this paper we address the important problem of optimizing regularization parameters in ...
In Gaussian mixture modeling, it is crucial to select the number of Gaussians or mixture model for a...
In many real-world applications, the amount of data available for training is often limited, and thu...
In this work we study and develop learning algorithms for networks based on regularization theory. I...
In this thesis, we present Regularized Learning with Feature Networks (RLFN), an approach for regula...
. In order to avoid overfitting in neural learning, a regularization term is added to the loss funct...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
Unsupervised neural networks, such as restricted Boltzmann machines (RBMs) and deep belief networks ...
The aim of this paper is to introduce two widely applicable regularization methods based on the dire...
Regularization of neural networks can alleviate overfitting in the training phase. Current regulariz...
In the present work we study In some areas, artificial feed forward neural networks are still a comp...
International audienceStudies on generalization performance of machine learning algorithms under the...
The large capacity of neural networks enables them to learn complex functions. To avoid overfitting,...
As for Gaussian mixture modeling, the key problem is to select the number of Gaussians in the mixtur...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Abstract. In this paper we address the important problem of optimizing regularization parameters in ...
In Gaussian mixture modeling, it is crucial to select the number of Gaussians or mixture model for a...
In many real-world applications, the amount of data available for training is often limited, and thu...
In this work we study and develop learning algorithms for networks based on regularization theory. I...
In this thesis, we present Regularized Learning with Feature Networks (RLFN), an approach for regula...
. In order to avoid overfitting in neural learning, a regularization term is added to the loss funct...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
Unsupervised neural networks, such as restricted Boltzmann machines (RBMs) and deep belief networks ...
The aim of this paper is to introduce two widely applicable regularization methods based on the dire...
Regularization of neural networks can alleviate overfitting in the training phase. Current regulariz...