This is the final version of the article. It first appeared from International Conference on Learning Representations via http://arxiv.org/abs/1511.05176v3Deep neural networks are powerful parametric models that can be trained efficiently using the backpropagation algorithm. Stochastic neural networks combine the power of large parametric functions with that of graphical models, which makes it possible to learn very complex distributions. However, as backpropagation is not directly applicable to stochastic networks that include discrete sampling operations within their computational graph, training such networks remains difficult. We present MuProp, an unbiased gradient estimator for stochastic networks, designed to make this task easier. M...
This paper demonstrates how the backpropagation algorithm (BP) and its variants can be accelerated s...
We marry ideas from deep neural networks and approximate Bayesian inference to derive a generalised ...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...
We introduce a novel training principle for prob-abilistic models that is an alternative to max-imum...
We introduce a novel training principle for probabilistic models that is an al-ternative to maximum ...
We introduce a novel training principle for probabilistic models that is an alternative to maximum l...
Large multilayer neural networks trained with backpropagation have recently achieved state-of-the-ar...
International audienceBackpropagating gradients through random variables is at the heart of numerous...
Multilayer Neural Networks (MNNs) are commonly trained using gradient descent-based methods, such as...
Despite of remarkable progress on deep learning, its hardware implementation beyond deep learning ac...
We investigate a new approach to compute the gradients of artificial neural networks (ANNs), based o...
In any data set there some of the data will be bad or noisy. This study identifies two types of nois...
This paper presents some simple techniques to improve the backpropagation algorithm. Since learning ...
We derive global H 1 optimal training algorithms for neural networks. These algorithms guarantee t...
Multilayer perceptrons (MLPs) or artificial neural nets are popular models used for non-linear regre...
This paper demonstrates how the backpropagation algorithm (BP) and its variants can be accelerated s...
We marry ideas from deep neural networks and approximate Bayesian inference to derive a generalised ...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...
We introduce a novel training principle for prob-abilistic models that is an alternative to max-imum...
We introduce a novel training principle for probabilistic models that is an al-ternative to maximum ...
We introduce a novel training principle for probabilistic models that is an alternative to maximum l...
Large multilayer neural networks trained with backpropagation have recently achieved state-of-the-ar...
International audienceBackpropagating gradients through random variables is at the heart of numerous...
Multilayer Neural Networks (MNNs) are commonly trained using gradient descent-based methods, such as...
Despite of remarkable progress on deep learning, its hardware implementation beyond deep learning ac...
We investigate a new approach to compute the gradients of artificial neural networks (ANNs), based o...
In any data set there some of the data will be bad or noisy. This study identifies two types of nois...
This paper presents some simple techniques to improve the backpropagation algorithm. Since learning ...
We derive global H 1 optimal training algorithms for neural networks. These algorithms guarantee t...
Multilayer perceptrons (MLPs) or artificial neural nets are popular models used for non-linear regre...
This paper demonstrates how the backpropagation algorithm (BP) and its variants can be accelerated s...
We marry ideas from deep neural networks and approximate Bayesian inference to derive a generalised ...
Stochastic binary hidden units in a multi-layer perceptron (MLP) network give at least three potenti...