The variational lower bound (a.k.a. ELBO or free energy) is the central objective for many learning algorithms including algorithms for deep unsupervised learning. Learning algorithms change model parameters such that the variational lower bound increases, and until the parameters are close to a stationary point of the learning dynamics. In this purely theoretical contribution, we show that (for a very large class of generative models) the variational lower bound is at all stationary points of learning equal to a sum of entropies. For models with one set of latents and one set observed variables, the sum consists of three entropies: (A) the (average) entropy of the variational distributions, (B) the negative entropy of the model's prior dis...
Variational inference (VI) is a popular method used within statistics and machine learning to approx...
International audienceMaximum entropy models provide the least constrained probability distributions...
We approach the subject of Statistical Mechanics from two different perspectives. In Part I we adopt...
The central objective function of a variational autoencoder (VAE) is its variational lower bound (th...
Variational inference is a technique for approximating intractable posterior distributions in order ...
Variational autoencoders (VAEs) are a popular framework for modeling complex data distributions; the...
Estimating information-theoretic quantities such as entropy and mutual information is central to man...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
We examine the minimization of information entropy for measures on the phase space of bounded domain...
We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can b...
In this work, we consider and analyze the sample complexity of model-free reinforcement learning wit...
The present paper elucidates a universal property of learning curves, which shows how the generaliza...
We study properties of popular near–uniform (Dirichlet) priors for learning undersampled probability...
The Variational AutoEncoder (VAE) learns simultaneously an inference and a generative model, but onl...
We present a latent variable generalisation of neural network softmax classification trained with cr...
Variational inference (VI) is a popular method used within statistics and machine learning to approx...
International audienceMaximum entropy models provide the least constrained probability distributions...
We approach the subject of Statistical Mechanics from two different perspectives. In Part I we adopt...
The central objective function of a variational autoencoder (VAE) is its variational lower bound (th...
Variational inference is a technique for approximating intractable posterior distributions in order ...
Variational autoencoders (VAEs) are a popular framework for modeling complex data distributions; the...
Estimating information-theoretic quantities such as entropy and mutual information is central to man...
The properties of flat minima in the empirical risk landscape of neural networks have been debated f...
We examine the minimization of information entropy for measures on the phase space of bounded domain...
We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can b...
In this work, we consider and analyze the sample complexity of model-free reinforcement learning wit...
The present paper elucidates a universal property of learning curves, which shows how the generaliza...
We study properties of popular near–uniform (Dirichlet) priors for learning undersampled probability...
The Variational AutoEncoder (VAE) learns simultaneously an inference and a generative model, but onl...
We present a latent variable generalisation of neural network softmax classification trained with cr...
Variational inference (VI) is a popular method used within statistics and machine learning to approx...
International audienceMaximum entropy models provide the least constrained probability distributions...
We approach the subject of Statistical Mechanics from two different perspectives. In Part I we adopt...