Multimodal variational autoencoders (VAEs) have shown promise as efficient generative models for weakly-supervised data. Yet, despite their advantage of weak supervision, they exhibit a gap in generative quality compared to unimodal VAEs, which are completely unsupervised. In an attempt to explain this gap, we uncover a fundamental limitation that applies to a large family of mixture-based multimodal VAEs. We prove that the sub-sampling of modalities enforces an undesirable upper bound on the multimodal ELBO and thereby limits the generative quality of the respective models. Empirically, we showcase the generative quality gap on both synthetic and real data and present the tradeoffs between different variants of multimodal VAEs. We find th...
Mixture models in variational inference (VI) is an active field of research. Recent works have estab...
Thesis (Master's)--University of Washington, 2022In this thesis, we conduct a thorough study of "Var...
Variational autoencoders (VAEs), as well as other generative models, have been shown to be efficient...
Multimodal VAEs have recently gained attention as efficient models for weakly-supervised generative ...
Multimodal Variational Autoencoders (VAEs) have been a subject of intense research in the past years...
A number of variational autoencoders (VAEs) have recently emerged with the aim of modeling multimoda...
Deep generative models often perform poorly in real-world applications due to the heterogeneity of n...
Multimodal learning is a framework for building models that make predictions based on different type...
A key advance in learning generative models is the use of amortized inference distributions that are...
Variational autoencoders (VAEs) is a strong family of deep generative models based on variational in...
In this paper we propose a conditional generative modelling (CGM) approach for unsupervised disentan...
Variational autoencoders (VAEs) have received considerable attention, since they allow us to learn e...
Recent work in unsupervised learning has focused on efficient inference and learning in latent varia...
Variational autoencoders (VAEs), as well as other generative models, have been shown to be efficient...
One of the major shortcomings of variational autoencoders is the inability to produce generations fr...
Mixture models in variational inference (VI) is an active field of research. Recent works have estab...
Thesis (Master's)--University of Washington, 2022In this thesis, we conduct a thorough study of "Var...
Variational autoencoders (VAEs), as well as other generative models, have been shown to be efficient...
Multimodal VAEs have recently gained attention as efficient models for weakly-supervised generative ...
Multimodal Variational Autoencoders (VAEs) have been a subject of intense research in the past years...
A number of variational autoencoders (VAEs) have recently emerged with the aim of modeling multimoda...
Deep generative models often perform poorly in real-world applications due to the heterogeneity of n...
Multimodal learning is a framework for building models that make predictions based on different type...
A key advance in learning generative models is the use of amortized inference distributions that are...
Variational autoencoders (VAEs) is a strong family of deep generative models based on variational in...
In this paper we propose a conditional generative modelling (CGM) approach for unsupervised disentan...
Variational autoencoders (VAEs) have received considerable attention, since they allow us to learn e...
Recent work in unsupervised learning has focused on efficient inference and learning in latent varia...
Variational autoencoders (VAEs), as well as other generative models, have been shown to be efficient...
One of the major shortcomings of variational autoencoders is the inability to produce generations fr...
Mixture models in variational inference (VI) is an active field of research. Recent works have estab...
Thesis (Master's)--University of Washington, 2022In this thesis, we conduct a thorough study of "Var...
Variational autoencoders (VAEs), as well as other generative models, have been shown to be efficient...