When using deep, multi-layered architectures to build generative models of data, it is difficult to train all layers at once. We propose a layer-wise training procedure admitting a performance guarantee compared to the global optimum. It is based on an optimistic proxy of future performance, the best latent marginal. We interpret auto-encoders in this setting as generative models, by showing that they train a lower bound of this criterion. We test the new learning procedure against a state of the art method (stacked RBMs), and find it to improve performance. Both theory and experiments highlight the importance, when training deep architectures, of using an inference model (from data to hidden variables) richer than the generative model (fro...
Advances in deep generative models are at the forefront of deep learning research because of the pro...
Since 2006, deep learning algorithms which rely on deep architectures with several layers of increas...
In the fields of natural language processing (NLP) and computer vision (CV), recent advances in gene...
When using deep, multi-layered architectures to build generative models of data, it is difficult to ...
Generative models are probabilistic models which aim at approximating the process by which a given d...
Complexity theory of circuits strongly suggests that deep architectures can be much more efcient (so...
2014 We give algorithms with provable guarantees that learn a class of deep nets in the generative m...
We introduce a deep, generative autoencoder ca-pable of learning hierarchies of distributed rep-rese...
Deep generative models allow us to learn hidden representations of data and generate new examples. T...
Deep learning provides us with ever-more-sophisticated neural networks that can be tuned via gradien...
We consider the problem of learning deep gener-ative models from data. We formulate a method that ge...
A deep generative model is characterized by a representation space, its distribution, and a neural n...
Building intelligent systems that are capable of extracting high-level representations from high-dim...
The problem Building good predictors on complex domains means learning complicated functions. These ...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Advances in deep generative models are at the forefront of deep learning research because of the pro...
Since 2006, deep learning algorithms which rely on deep architectures with several layers of increas...
In the fields of natural language processing (NLP) and computer vision (CV), recent advances in gene...
When using deep, multi-layered architectures to build generative models of data, it is difficult to ...
Generative models are probabilistic models which aim at approximating the process by which a given d...
Complexity theory of circuits strongly suggests that deep architectures can be much more efcient (so...
2014 We give algorithms with provable guarantees that learn a class of deep nets in the generative m...
We introduce a deep, generative autoencoder ca-pable of learning hierarchies of distributed rep-rese...
Deep generative models allow us to learn hidden representations of data and generate new examples. T...
Deep learning provides us with ever-more-sophisticated neural networks that can be tuned via gradien...
We consider the problem of learning deep gener-ative models from data. We formulate a method that ge...
A deep generative model is characterized by a representation space, its distribution, and a neural n...
Building intelligent systems that are capable of extracting high-level representations from high-dim...
The problem Building good predictors on complex domains means learning complicated functions. These ...
Understanding the impact of data structure on the computational tractability of learning is a key ch...
Advances in deep generative models are at the forefront of deep learning research because of the pro...
Since 2006, deep learning algorithms which rely on deep architectures with several layers of increas...
In the fields of natural language processing (NLP) and computer vision (CV), recent advances in gene...