We develop a new component analysis framework, the Noisy-Or Component Analyzer (NOCA), that targets high-dimensional binary data. NOCA is a probabilistic latent variable model that assumes the expression of observed high-dimensional binary data is driven by a small number of hidden binary sources combined via noisy-or units. The component analysis procedure is equivalent to learning of NOCA parameters. Since the classical EM formulation of the NOCA learning problem is intractable, we develop its variational approximation. We test the NOCA framework on two problems: (1) a synthetic image-decomposition problem and (2) a co-citation data analysis problem for thousands of CiteSeer documents. We demonstrate good performance of the new mo...
In a latent variable model, an overcomplete representation is one in which the number of latent vari...
Inference of the document-specific topic distributions in latent Dirichlet allocation (LDA) [2] and ...
Summarising a high dimensional data set with a low dimensional embedding is a standard approach for ...
We develop a new component analysis framework, the Noisy-Or Component Analyzer (NOCA), that targets...
We propose a method based on the probabilistic latent componentanalysis (PLCA) in which we use expon...
In this paper we present a model which can decompose a probability densities or count data into a se...
In bioinformatics it is often desirable to combine data from various measurement sources and thus st...
International audienceWe present a Bayesian approach for Sparse Component Analysis (SCA) in the nois...
[[abstract]]Most independent component analysis methods for blind source separation rely on the fund...
We propose a probabilistic model based on Independent Component Analysis for learning multiple relat...
Component Analysis (CA) consists of a set of statistical techniques that decompose data to appropria...
We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of ...
We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of d...
We present a general framework for data analysis and visualisation by means of topographic organizat...
Mixtures of probabilistic principal component analyzers model high-dimensional nonlinear data by com...
In a latent variable model, an overcomplete representation is one in which the number of latent vari...
Inference of the document-specific topic distributions in latent Dirichlet allocation (LDA) [2] and ...
Summarising a high dimensional data set with a low dimensional embedding is a standard approach for ...
We develop a new component analysis framework, the Noisy-Or Component Analyzer (NOCA), that targets...
We propose a method based on the probabilistic latent componentanalysis (PLCA) in which we use expon...
In this paper we present a model which can decompose a probability densities or count data into a se...
In bioinformatics it is often desirable to combine data from various measurement sources and thus st...
International audienceWe present a Bayesian approach for Sparse Component Analysis (SCA) in the nois...
[[abstract]]Most independent component analysis methods for blind source separation rely on the fund...
We propose a probabilistic model based on Independent Component Analysis for learning multiple relat...
Component Analysis (CA) consists of a set of statistical techniques that decompose data to appropria...
We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of ...
We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of d...
We present a general framework for data analysis and visualisation by means of topographic organizat...
Mixtures of probabilistic principal component analyzers model high-dimensional nonlinear data by com...
In a latent variable model, an overcomplete representation is one in which the number of latent vari...
Inference of the document-specific topic distributions in latent Dirichlet allocation (LDA) [2] and ...
Summarising a high dimensional data set with a low dimensional embedding is a standard approach for ...