Helmholtz Machines (HMs) are a class of generative models composed of two Sigmoid Belief Networks (SBNs), acting respectively as an encoder and a decoder. These models are commonly trained using a two-step optimization algorithm called Wake-Sleep (WS) and more recently by improved versions, such as Reweighted Wake-Sleep (RWS) and Bidirectional Helmholtz Machines (BiHM). The locality of the connections in an SBN induces sparsity in the Fisher Information Matrices associated to the probabilistic models, in the form of a finely-grained block-diagonal structure. In this paper we exploit this property to efficiently train SBNs and HMs using the natural gradient. We present a novel algorithm, called Natural Reweighted Wake-Sleep (NRWS), that corr...
This paper presents a novel natural gradient and Hessian-free (NGHF) optimisation framework for neur...
Unsupervised learning is largely concerned with finding structure among sets of input patterns such ...
AbstractLearning real weights for a McCulloch-Pitts neuron is equivalent to linear programming and c...
We describe the "wake-sleep'' algorithm that allows a multilayer, unsupervised, stochastic neural ne...
When a parameter space has a certain underlying structure, the ordinary gradient of a function does ...
The wake-sleep algorithm (Hinton, Dayan, Frey and Neal 1995) is a relatively efficient method of fit...
A b s t r a c t. Recently, several evolutionary algorithms have been proposed that build and use an ...
Neural network training algorithms have always suffered from the problem of local minima. The advent...
Understanding intelligence and how it allows humans to learn, to make decision and form memories, is...
We introduce a new approach to learning in hierarchical latent-variable generative models called the...
We analyse natural gradient learning in a two-layer feed-forward neural network using a statistical ...
Neural network models able to approximate and sample high-dimensional probability distributions are ...
Deemed as the third generation of neural networks, the event-driven Spiking Neural Networks(SNNs) co...
Natural gradient descent (NGD) is an on-line algorithm for redefining the steepest descent direction...
This paper sums up the main contributions of the PhD Dissertation with an homonymous name to the cur...
This paper presents a novel natural gradient and Hessian-free (NGHF) optimisation framework for neur...
Unsupervised learning is largely concerned with finding structure among sets of input patterns such ...
AbstractLearning real weights for a McCulloch-Pitts neuron is equivalent to linear programming and c...
We describe the "wake-sleep'' algorithm that allows a multilayer, unsupervised, stochastic neural ne...
When a parameter space has a certain underlying structure, the ordinary gradient of a function does ...
The wake-sleep algorithm (Hinton, Dayan, Frey and Neal 1995) is a relatively efficient method of fit...
A b s t r a c t. Recently, several evolutionary algorithms have been proposed that build and use an ...
Neural network training algorithms have always suffered from the problem of local minima. The advent...
Understanding intelligence and how it allows humans to learn, to make decision and form memories, is...
We introduce a new approach to learning in hierarchical latent-variable generative models called the...
We analyse natural gradient learning in a two-layer feed-forward neural network using a statistical ...
Neural network models able to approximate and sample high-dimensional probability distributions are ...
Deemed as the third generation of neural networks, the event-driven Spiking Neural Networks(SNNs) co...
Natural gradient descent (NGD) is an on-line algorithm for redefining the steepest descent direction...
This paper sums up the main contributions of the PhD Dissertation with an homonymous name to the cur...
This paper presents a novel natural gradient and Hessian-free (NGHF) optimisation framework for neur...
Unsupervised learning is largely concerned with finding structure among sets of input patterns such ...
AbstractLearning real weights for a McCulloch-Pitts neuron is equivalent to linear programming and c...