In this work, we propose to progressively increase the training difficulty during learning a neural network model via a novel strategy which we call mini-batch trimming. This strategy makes sure that the optimizer puts its focus in the later training stages on the more difficult samples, which we identify as the ones with the highest loss in the current mini-batch. The strategy is very easy to integrate into an existing training pipeline and does not necessitate a change of the network model. Experiments on several image classification problems show that mini-batch trimming is able to increase the generalization ability (measured via final test error) of the trained model
Neural Networks (NN) can be trained to perform tasks such as image and handwriting recognition, cred...
Deep neural networks (DNNs) require large amounts of labeled data for model training. However, label...
© 2018. The copyright of this document resides with its authors. In this paper, we propose a simple ...
We present a comprehensive framework of search methods, such as simulated annealing and batch traini...
Training Deep Neural Networks is complicated by the fact that the distribution of each layer’s input...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
Large-batch SGD is important for scaling training of deep neural networks. However, without fine-tun...
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via stat...
Neural Networks (NN) can be trained to perform tasks such as image and handwriting recognition, cred...
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via stat...
We propose a metric for evaluating the generalization ability of deep neural networks trained with m...
During minibatch gradient-based optimization, the contribution of observations to the updating of th...
We introduce a new approach to the training of classifiers for performance on multiple tasks. The pr...
Neural Networks (NN) can be trained to perform tasks such as image and handwriting recognition, cred...
Deep neural networks (DNNs) require large amounts of labeled data for model training. However, label...
© 2018. The copyright of this document resides with its authors. In this paper, we propose a simple ...
We present a comprehensive framework of search methods, such as simulated annealing and batch traini...
Training Deep Neural Networks is complicated by the fact that the distribution of each layer’s input...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
In the last decade, motivated by the success of Deep Learning, the scientific community proposed sev...
Large-batch SGD is important for scaling training of deep neural networks. However, without fine-tun...
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via stat...
Neural Networks (NN) can be trained to perform tasks such as image and handwriting recognition, cred...
Batch Normalization (BN) (Ioffe and Szegedy 2015) normalizes the features of an input image via stat...
We propose a metric for evaluating the generalization ability of deep neural networks trained with m...
During minibatch gradient-based optimization, the contribution of observations to the updating of th...
We introduce a new approach to the training of classifiers for performance on multiple tasks. The pr...
Neural Networks (NN) can be trained to perform tasks such as image and handwriting recognition, cred...
Deep neural networks (DNNs) require large amounts of labeled data for model training. However, label...
© 2018. The copyright of this document resides with its authors. In this paper, we propose a simple ...