to be presented at ICML2022 in Baltimore, MDInternational audienceMutual Information (MI) has been widely used as a loss regularizer for training neural networks. This has been particularly effective when learn disentangled or compressed representations of high dimensional data. However, differential entropy (DE), another fundamental measure of information, has not found widespread use in neural network training. Although DE offers a potentially wider range of applications than MI, off-the-shelf DE estimators are either non differentiable, computationally intractable or fail to adapt to changes in the underlying distribution. These drawbacks prevent them from being used as regularizers in neural networks training. To address shortcomings in...
To train robust deep neural networks (DNNs), we systematically study several target modification app...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
Mutual Information (MI) has been widely used as a loss regularizer for training neural networks. Thi...
Neural networks have dramatically increased our capacity to learn from large, high-dimensional datas...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
We examine a class of stochastic deep learning models with a tractable method to compute information...
We examine a class of stochastic deep learning models with a tractable method to compute information...
International audienceStudies on generalization performance of machine learning algorithms under the...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Generative Adversarial Networks (GANs) have shown compelling results in various tasks and applicatio...
Training Artificial Neural Networks (ANNs) is a non-trivial task. In the last years, there has been ...
When humans learn a new concept, they might ignore examples that they cannot make sense of at first,...
To train robust deep neural networks (DNNs), we systematically study several target modification app...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
Mutual Information (MI) has been widely used as a loss regularizer for training neural networks. Thi...
Neural networks have dramatically increased our capacity to learn from large, high-dimensional datas...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
International audienceWe examine a class of stochastic deep learning models with a tractable method ...
We examine a class of stochastic deep learning models with a tractable method to compute information...
We examine a class of stochastic deep learning models with a tractable method to compute information...
International audienceStudies on generalization performance of machine learning algorithms under the...
Deep neural language models like GPT-2 is undoubtedly strong at text generation, but often requires ...
Generative Adversarial Networks (GANs) have shown compelling results in various tasks and applicatio...
Training Artificial Neural Networks (ANNs) is a non-trivial task. In the last years, there has been ...
When humans learn a new concept, they might ignore examples that they cannot make sense of at first,...
To train robust deep neural networks (DNNs), we systematically study several target modification app...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...