Each parameter ` in an abstract parameter space \Theta is associated with a different probability distribution on a set Y . A parameter ` is chosen at random from \Theta according to some a priori distribution on \Theta, and n conditionally independent random variables Y n = Y 1 ; : : : Yn are observed with common distribution determined by `. We obtain bounds on the mutual information between the random variable \Theta, giving the choice of parameter, and the random variable Y n , giving the sequence of observations. We also bound the supremum of the mutual information, over choices of the prior distribution on \Theta. These quantities have applications in density estimation, computational learning theory, universal coding, hypothesi...