In this paper we introduce an algorithm for learning hybrid Bayesian networks from data. The result of the algorithm is a network where the conditional distribution for each variable is a mixture of truncated exponentials (MTE), so that no restrictions on the network topology are imposed. The structure of the network is obtained by searching over the space of candidate networks using optimisation methods. The conditional densities are estimated by means of Gaussian kernel densities that afterwards are approximated by MTEs, so that the resulting network is appropriate for using standard algorithms for probabilistic reasoning. The behaviour of the proposed algorithm is tested using a set of real-world and arti cially generated databases
AbstractThe main goal of this paper is to describe inference in hybrid Bayesian networks (BNs) using...
Bayesian networks with mixtures of truncated exponentials (MTEs) support efficient in-ference algori...
Hybrid Bayesian networks efficiently encode a joint probability distribution over a set of continuou...
In this paper we introduce an algorithm for learning hybrid Bayesian networks from data. The result ...
AbstractIn this paper we introduce an algorithm for learning hybrid Bayesian networks from data. The...
AbstractMixtures of truncated exponentials (MTE) potentials are an alternative to discretization for...
The main goal of this paper is to describe a method for exact inference in general hybrid Bayesian n...
In this paper we introduce a hill-climbing algorithm for structural learning of Bayesian networks fr...
The MTE (Mixture of Truncated Exponentials) model allows to deal with Bayesian networks containing d...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization for solving...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization and Monte C...
In this paper we investigate methods for learning hybrid Bayesian networks from data. First we utili...
Has been accepted for publication in the International Journal of Approximate Reasoning, Elsevier Sc...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization and Monte C...
In this paper, the first algorithm for learning hybrid Bayesian Networks with Gaussian mixture and D...
AbstractThe main goal of this paper is to describe inference in hybrid Bayesian networks (BNs) using...
Bayesian networks with mixtures of truncated exponentials (MTEs) support efficient in-ference algori...
Hybrid Bayesian networks efficiently encode a joint probability distribution over a set of continuou...
In this paper we introduce an algorithm for learning hybrid Bayesian networks from data. The result ...
AbstractIn this paper we introduce an algorithm for learning hybrid Bayesian networks from data. The...
AbstractMixtures of truncated exponentials (MTE) potentials are an alternative to discretization for...
The main goal of this paper is to describe a method for exact inference in general hybrid Bayesian n...
In this paper we introduce a hill-climbing algorithm for structural learning of Bayesian networks fr...
The MTE (Mixture of Truncated Exponentials) model allows to deal with Bayesian networks containing d...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization for solving...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization and Monte C...
In this paper we investigate methods for learning hybrid Bayesian networks from data. First we utili...
Has been accepted for publication in the International Journal of Approximate Reasoning, Elsevier Sc...
Mixtures of truncated exponentials (MTE) potentials are an alternative to discretization and Monte C...
In this paper, the first algorithm for learning hybrid Bayesian Networks with Gaussian mixture and D...
AbstractThe main goal of this paper is to describe inference in hybrid Bayesian networks (BNs) using...
Bayesian networks with mixtures of truncated exponentials (MTEs) support efficient in-ference algori...
Hybrid Bayesian networks efficiently encode a joint probability distribution over a set of continuou...