We consider the analysis of probability distributions through their associated covariance operators from reproducing kernel Hilbert spaces. We show that the von Neumann entropy and relative entropy of these operators are intimately related to the usual notions of Shannon entropy and relative entropy, and share many of their properties. They come together with efficient estimation algorithms from various oracles on the probability distributions. We also consider product spaces and show that for tensor product kernels, we can define notions of mutual information and joint entropies, which can then characterize independence perfectly, but only partially conditional independence. We finally show how these new notions of relative entropy lead to...
Abstract—Upper and lower bounds are obtained for the joint entropy of a collection of random variabl...
We review a decision theoretic, i.e., utility-based, motivation for entropy and Kullback-Leibler rel...
Axiomatic characterizations of Shannon entropy, Kullback I-divergence, and some generalized informat...
We consider the analysis of probability distributions through their associated covariance operators ...
While kernel methods are the basis of many popular techniques in supervised learning, they are less ...
It is well known that in Information Theory and Machine Learning the Kullback-Leibler divergence, wh...
We introduce an axiomatic approach to entropies and relative entropies that relies only on minimal i...
Entropy and conditional mutual information are the key quantities information theory provides to mea...
The role of kernels is central to machine learning. Motivated by the importance of power-law distrib...
Abstract—The role of kernels is central to machine learning. Motivated by the importance of power-la...
In a probability space, the partition fiber relative to a probability vector v is the set of all ord...
We introduce two new functionals, the constrained covariance and the kernel mutual information, to m...
We analyze entropic uncertainty relations for two orthogonal measurements on a N-dimensional Hilbert...
We consider Bayesian estimation of information-theoretic quantities from data, using a Dirichlet pr...
This paper is a review of a particular approach to the method of maximum entropy as a general framew...
Abstract—Upper and lower bounds are obtained for the joint entropy of a collection of random variabl...
We review a decision theoretic, i.e., utility-based, motivation for entropy and Kullback-Leibler rel...
Axiomatic characterizations of Shannon entropy, Kullback I-divergence, and some generalized informat...
We consider the analysis of probability distributions through their associated covariance operators ...
While kernel methods are the basis of many popular techniques in supervised learning, they are less ...
It is well known that in Information Theory and Machine Learning the Kullback-Leibler divergence, wh...
We introduce an axiomatic approach to entropies and relative entropies that relies only on minimal i...
Entropy and conditional mutual information are the key quantities information theory provides to mea...
The role of kernels is central to machine learning. Motivated by the importance of power-law distrib...
Abstract—The role of kernels is central to machine learning. Motivated by the importance of power-la...
In a probability space, the partition fiber relative to a probability vector v is the set of all ord...
We introduce two new functionals, the constrained covariance and the kernel mutual information, to m...
We analyze entropic uncertainty relations for two orthogonal measurements on a N-dimensional Hilbert...
We consider Bayesian estimation of information-theoretic quantities from data, using a Dirichlet pr...
This paper is a review of a particular approach to the method of maximum entropy as a general framew...
Abstract—Upper and lower bounds are obtained for the joint entropy of a collection of random variabl...
We review a decision theoretic, i.e., utility-based, motivation for entropy and Kullback-Leibler rel...
Axiomatic characterizations of Shannon entropy, Kullback I-divergence, and some generalized informat...