We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can be detrimental to the process of learning an inference network by reducing the signal-to-noise ratio of the gradient estimator. Our results call into question common implicit assumptions that tighter ELBOs are better variational objectives for simultaneous model learning and inference amortization schemes. Based on our insights, we introduce three new algorithms: the partially importance weighted auto-encoder (PIWAE), the multiply importance weighted auto-encoder (MIWAE), and the combination importance weighted autoencoder (CIWAE), each of which includes the standard importance weighted auto-encoder (IWAE) as a special case. We show that each ...
Inference models are a key component in scaling variational inference to deep latent variable models...
In the past few years Generative models have become an interesting topic in the field of Machine Lea...
How can we perform efficient inference and learning in directed probabilistic models, in the presenc...
A key advance in learning generative models is the use of amortized inference distributions that are...
A deep latent variable model is a powerful tool for modelling complex distributions. However, in ord...
Deep generative models allow us to learn hidden representations of data and generate new examples. T...
Semi-supervised variational autoencoders (VAEs) have obtained strong results, but have also encounte...
Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at th...
The Variational AutoEncoder (VAE) learns simultaneously an inference and a generative model, but onl...
This version of the article has been accepted for publication, after peer review (when applicable) a...
We revisit the theory of importance weighted variational inference (IWVI), a promising strategy for ...
The central objective function of a variational autoencoder (VAE) is its variational lower bound (th...
Thesis (Master's)--University of Washington, 2022In this thesis, we conduct a thorough study of "Var...
We claim that a source of severe failures for Variational Auto-Encoders is the choice of the distrib...
Variational auto-encoders (VAE) are popular deep latent variable models which are trained by maximiz...
Inference models are a key component in scaling variational inference to deep latent variable models...
In the past few years Generative models have become an interesting topic in the field of Machine Lea...
How can we perform efficient inference and learning in directed probabilistic models, in the presenc...
A key advance in learning generative models is the use of amortized inference distributions that are...
A deep latent variable model is a powerful tool for modelling complex distributions. However, in ord...
Deep generative models allow us to learn hidden representations of data and generate new examples. T...
Semi-supervised variational autoencoders (VAEs) have obtained strong results, but have also encounte...
Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at th...
The Variational AutoEncoder (VAE) learns simultaneously an inference and a generative model, but onl...
This version of the article has been accepted for publication, after peer review (when applicable) a...
We revisit the theory of importance weighted variational inference (IWVI), a promising strategy for ...
The central objective function of a variational autoencoder (VAE) is its variational lower bound (th...
Thesis (Master's)--University of Washington, 2022In this thesis, we conduct a thorough study of "Var...
We claim that a source of severe failures for Variational Auto-Encoders is the choice of the distrib...
Variational auto-encoders (VAE) are popular deep latent variable models which are trained by maximiz...
Inference models are a key component in scaling variational inference to deep latent variable models...
In the past few years Generative models have become an interesting topic in the field of Machine Lea...
How can we perform efficient inference and learning in directed probabilistic models, in the presenc...