The vicinal risk minimization (VRM) principle is an empirical risk minimization (ERM) variant that replaces Dirac masses with vicinal functions. There is strong numerical and theoretical evidence showing that VRM outperforms ERM in terms of generalization if appropriate vicinal functions are chosen. Mixup Training (MT), a popular choice of vicinal distribution, improves generalization performance of models by introducing globally linear behavior in between training examples. Apart from generalization, recent works have shown that mixup trained models are relatively robust to input perturbations/corruptions and at same time are calibrated better than their non-mixup counterparts. In this work, we investigate the benefits of defining these vi...
Deep neural networks excel at solving intuitive tasks that are hard to describe formally, such as cl...
The Mixup scheme suggests mixing a pair of samples to create an augmented training sample and has ga...
The theoretical and empirical performance of Empirical Risk Minimization (ERM) often suffers when lo...
The Vicinal Risk Minimization principle establishes a bridge between generative models and methods d...
Mixup is a data augmentation technique that creates new examples as convex combinations of training ...
We propose and motivate the use of vicinal-risk minimization (VRM) for training genetic programming ...
We propose the use of Vapnik's vicinal risk minimization (VRM) for training decision trees to approx...
Mixup is a popular data augmentation technique for training deep neural networks where additional sa...
International audienceDeployed in the context of supervised learning, Mixup is a data-dependent regu...
Mixture models in variational inference (VI) is an active field of research. Recent works have estab...
International audienceDeployed in the context of supervised learning, Mixup is a data-dependent regu...
MixUp (Zhang et al. 2017) is a recently proposed dataaugmentation scheme, which linearly interpolate...
Mixup is a recently proposed learning paradigm that improves the generalization of deep neural netwo...
Mixup is an efficient data augmentation approach that improves the generalization of neural networks...
Mixup is a popular regularization technique for training deep neural networks that can improve gene...
Deep neural networks excel at solving intuitive tasks that are hard to describe formally, such as cl...
The Mixup scheme suggests mixing a pair of samples to create an augmented training sample and has ga...
The theoretical and empirical performance of Empirical Risk Minimization (ERM) often suffers when lo...
The Vicinal Risk Minimization principle establishes a bridge between generative models and methods d...
Mixup is a data augmentation technique that creates new examples as convex combinations of training ...
We propose and motivate the use of vicinal-risk minimization (VRM) for training genetic programming ...
We propose the use of Vapnik's vicinal risk minimization (VRM) for training decision trees to approx...
Mixup is a popular data augmentation technique for training deep neural networks where additional sa...
International audienceDeployed in the context of supervised learning, Mixup is a data-dependent regu...
Mixture models in variational inference (VI) is an active field of research. Recent works have estab...
International audienceDeployed in the context of supervised learning, Mixup is a data-dependent regu...
MixUp (Zhang et al. 2017) is a recently proposed dataaugmentation scheme, which linearly interpolate...
Mixup is a recently proposed learning paradigm that improves the generalization of deep neural netwo...
Mixup is an efficient data augmentation approach that improves the generalization of neural networks...
Mixup is a popular regularization technique for training deep neural networks that can improve gene...
Deep neural networks excel at solving intuitive tasks that are hard to describe formally, such as cl...
The Mixup scheme suggests mixing a pair of samples to create an augmented training sample and has ga...
The theoretical and empirical performance of Empirical Risk Minimization (ERM) often suffers when lo...