<div><p>The use of mutual information as a similarity measure in agglomerative hierarchical clustering (AHC) raises an important issue: some correction needs to be applied for the dimensionality of variables. In this work, we formulate the decision of merging dependent multivariate normal variables in an AHC procedure as a Bayesian model comparison. We found that the Bayesian formulation naturally shrinks the empirical covariance matrix towards a matrix set a priori (e.g., the identity), provides an automated stopping rule, and corrects for dimensionality using a term that scales up the measure as a function of the dimensionality of the variables. Also, the resulting log Bayes factor is asymptotically proportional to the plug-in estimate of...
Clustering analysis is an important tool in studying gene expression data. The Bayesian hierarchical...
Current methods for hierarchical clustering of data either operate on features of the data or make l...
Mutual information is a very popular measure for comparing clusterings. Previous work has shown that...
The use of mutual information as a similarity measure in agglomerative hierarchical clustering (AHC)...
International audienceThe use of mutual information as a similarity measure in agglomerative hierarc...
International audienceThe use of mutual information as a similarity measure in agglomerative hierarc...
The use of mutual information as a similarity measure in agglomerative hierarchical cluster-ing (AHC...
We present a novel algorithm for agglomerative hierarchical clustering based on evaluating marginal ...
We present a conceptually simple method for hierarchical clustering of data called mutual informatio...
Most clustering algorithms assume that all dimensions of the data can be described by a single struc...
Clustering techniques aim organizing data into groups whose members are similar. A key element of th...
Clustering techniques aim organizing data into groups whose members are similar. A key element of th...
Clustering by maximizing the dependency between (margin) groupings or partitionings of co-occurring...
<p>Clustering methods are designed to separate heterogeneous data into groups of similar objects suc...
Hierarchical models are extensively used in pharmacokinetics and longitudinal studies. When the esti...
Clustering analysis is an important tool in studying gene expression data. The Bayesian hierarchical...
Current methods for hierarchical clustering of data either operate on features of the data or make l...
Mutual information is a very popular measure for comparing clusterings. Previous work has shown that...
The use of mutual information as a similarity measure in agglomerative hierarchical clustering (AHC)...
International audienceThe use of mutual information as a similarity measure in agglomerative hierarc...
International audienceThe use of mutual information as a similarity measure in agglomerative hierarc...
The use of mutual information as a similarity measure in agglomerative hierarchical cluster-ing (AHC...
We present a novel algorithm for agglomerative hierarchical clustering based on evaluating marginal ...
We present a conceptually simple method for hierarchical clustering of data called mutual informatio...
Most clustering algorithms assume that all dimensions of the data can be described by a single struc...
Clustering techniques aim organizing data into groups whose members are similar. A key element of th...
Clustering techniques aim organizing data into groups whose members are similar. A key element of th...
Clustering by maximizing the dependency between (margin) groupings or partitionings of co-occurring...
<p>Clustering methods are designed to separate heterogeneous data into groups of similar objects suc...
Hierarchical models are extensively used in pharmacokinetics and longitudinal studies. When the esti...
Clustering analysis is an important tool in studying gene expression data. The Bayesian hierarchical...
Current methods for hierarchical clustering of data either operate on features of the data or make l...
Mutual information is a very popular measure for comparing clusterings. Previous work has shown that...