Recently, we introduced a simple variational bound on mutual information, that resolves some of the difficulties in the application of information theory to machine learning. Here we study a specific application to Gaussian channels. It is well known that PCA may be viewed as the solution to maximizing information transmission between a high dimensional vector and its low dimensional representation . However, such results are based on assumptions of Gaussianity of the sources. In this paper, we show how our mutual information bound, when applied to this arena, gives PCA solutions, without the need for the Gaussian assumption. Furthermore, it naturally generalizes to providing an objective function for Kernel PCA, enabling the principled sel...
Abstract—Let X,Y,Z be zero-mean, jointly Gaussian ran-dom vectors of dimensions nx, ny and nz, respe...
summary:We investigate the sets of joint probability distributions that maximize the average multi-i...
This thesis introduces the Mutual Information Machine (MIM), an autoencoder model for learning j...
Information maximization is a common framework of unsupervised learning, which may be used for extr...
The goal of neural processing assemblies is varied, and in many cases still rather unclear. However,...
AbstractFor the Gaussian channel Y(t) = Φ(ξ(s), Y(s); s ≦ t) + X(t), the mutual information I(ξ, Y) ...
In this paper, derivatives of mutual information for a general linear Gaussian vector channel are co...
We consider a linear, one-layer feedforward neural network performing a coding task. The goal of the...
We consider the estimation of a signal from the knowledge of its noisy linear random Gaussian projec...
AbstractIn the Gaussian channel Y(t) = Φ(t) + X(t) = message + noise, where Φ(t) and X(t) are mutual...
Successful applications of InfoNCE and its variants have popularized the use of contrastive variatio...
Variational inference with a factorized Gaussian posterior estimate is a widely-used approach for le...
This paper was accepted for publication to Machine Learning (Springer). Overfitting data is a well-k...
We study the design of linear precoders for maximization of the mutual information in MIMO systems ...
We derive a tight lower bound on equivocation (conditional entropy), or equivalently a tight upper b...
Abstract—Let X,Y,Z be zero-mean, jointly Gaussian ran-dom vectors of dimensions nx, ny and nz, respe...
summary:We investigate the sets of joint probability distributions that maximize the average multi-i...
This thesis introduces the Mutual Information Machine (MIM), an autoencoder model for learning j...
Information maximization is a common framework of unsupervised learning, which may be used for extr...
The goal of neural processing assemblies is varied, and in many cases still rather unclear. However,...
AbstractFor the Gaussian channel Y(t) = Φ(ξ(s), Y(s); s ≦ t) + X(t), the mutual information I(ξ, Y) ...
In this paper, derivatives of mutual information for a general linear Gaussian vector channel are co...
We consider a linear, one-layer feedforward neural network performing a coding task. The goal of the...
We consider the estimation of a signal from the knowledge of its noisy linear random Gaussian projec...
AbstractIn the Gaussian channel Y(t) = Φ(t) + X(t) = message + noise, where Φ(t) and X(t) are mutual...
Successful applications of InfoNCE and its variants have popularized the use of contrastive variatio...
Variational inference with a factorized Gaussian posterior estimate is a widely-used approach for le...
This paper was accepted for publication to Machine Learning (Springer). Overfitting data is a well-k...
We study the design of linear precoders for maximization of the mutual information in MIMO systems ...
We derive a tight lower bound on equivocation (conditional entropy), or equivalently a tight upper b...
Abstract—Let X,Y,Z be zero-mean, jointly Gaussian ran-dom vectors of dimensions nx, ny and nz, respe...
summary:We investigate the sets of joint probability distributions that maximize the average multi-i...
This thesis introduces the Mutual Information Machine (MIM), an autoencoder model for learning j...