© 1979-2012 IEEE. Recent years have witnessed the success of deep neural networks in dealing with a plenty of practical problems. Dropout has played an essential role in many successful deep neural networks, by inducing regularization in the model training. In this paper, we present a new regularized training approach: Shakeout. Instead of randomly discarding units as Dropout does at the training stage, Shakeout randomly chooses to enhance or reverse each unit's contribution to the next layer. This minor modification of Dropout has the statistical trait: the regularizer induced by Shakeout adaptively combines L-{0} , L-{1} and L-{2} regularization terms. Our classification experiments with representative deep architectures on image datasets...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...
Adversarial training has been shown to regularize deep neural networks in addition to increasing the...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...
© Copyright 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rig...
Recent years have witnessed the success of deep neural networks in dealing with a plenty of practica...
Regularization is essential when training large neural networks. As deep neural networks can be math...
The undeniable computational power of artificial neural networks has granted the scientific communit...
Deep neural nets with a large number of parameters are very powerful machine learning systems. Howev...
Deep neural networks often consist of a great number of trainable parameters for extracting powerful...
Deep neural networks often consist of a great number of trainable parameters for extracting powerful...
Recently it has been shown that when training neural networks on a limited amount of data, randomly ...
Dropout is one of the most popular regularization methods used in deep learning. The general form of...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
We introduce DropConnect, a generalization of Dropout (Hinton et al., 2012), for regular-izing large...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...
Adversarial training has been shown to regularize deep neural networks in addition to increasing the...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...
© Copyright 2016, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rig...
Recent years have witnessed the success of deep neural networks in dealing with a plenty of practica...
Regularization is essential when training large neural networks. As deep neural networks can be math...
The undeniable computational power of artificial neural networks has granted the scientific communit...
Deep neural nets with a large number of parameters are very powerful machine learning systems. Howev...
Deep neural networks often consist of a great number of trainable parameters for extracting powerful...
Deep neural networks often consist of a great number of trainable parameters for extracting powerful...
Recently it has been shown that when training neural networks on a limited amount of data, randomly ...
Dropout is one of the most popular regularization methods used in deep learning. The general form of...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
Despite powerful representation ability, deep neural networks (DNNs) are prone to over-fitting, beca...
We introduce DropConnect, a generalization of Dropout (Hinton et al., 2012), for regular-izing large...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...
Adversarial training has been shown to regularize deep neural networks in addition to increasing the...
Recently, it was shown that deep neural networks perform very well if the activities of hidden units...