Modelers use automatic differentiation (AD) of computation graphs to implement complex deep learning models without defining gradient computations. Stochastic AD extends AD to stochastic computation graphs with sampling steps, which arise when modelers handle the intractable expectations common in reinforcement learning and variational inference. However, current methods for stochastic AD are limited: They are either only applicable to continuous random variables and differentiable functions, or can only use simple but high variance score-function estimators. To overcome these limitations, we introduce Storchastic, a new framework for AD of stochastic computation graphs. Storchastic allows the modeler to choose from a wide variety of gradie...
We introduce the stochastic gradient descent algorithm used in the computational network toolkit (CN...
International audienceStochastic approximation (SA) is a classical algorithm that has had since the ...
This thesis presents five contributions to machine learning, with themes of differentiability and Ba...
By enabling correct differentiation in Stochastic Computation Graphs (SCGs), the infinitely differen...
Optimizing via stochastic gradients is a powerful and exible technique ubiquitously used in machine ...
By enabling correct differentiation in stochastic computation graphs (SCGs), the infinitely differe...
237 pagesIt seems that in the current age, computers, computation, and data have an increasingly imp...
The reparameterization trick enables optimizing large scale stochastic computation graphs via gradie...
Derivatives, mostly in the form of gradients and Hessians, are ubiquitous in machine learning. Autom...
In the age of artificial intelligence, the best approach to handling huge amounts of data is a treme...
The goal of this paper is to debunk and dispel the magic behind black-box optimizers and stochastic ...
Discrete expectations arise in various machine learning tasks, and we often need to backpropagate th...
Automatic differentiation, as implemented today, does not have a simple mathematical model adapted t...
The score function estimator is widely used for estimating gradients of stochastic objectives in sto...
The deep learning community has devised a diverse set of methods to make gradient optimization, usin...
We introduce the stochastic gradient descent algorithm used in the computational network toolkit (CN...
International audienceStochastic approximation (SA) is a classical algorithm that has had since the ...
This thesis presents five contributions to machine learning, with themes of differentiability and Ba...
By enabling correct differentiation in Stochastic Computation Graphs (SCGs), the infinitely differen...
Optimizing via stochastic gradients is a powerful and exible technique ubiquitously used in machine ...
By enabling correct differentiation in stochastic computation graphs (SCGs), the infinitely differe...
237 pagesIt seems that in the current age, computers, computation, and data have an increasingly imp...
The reparameterization trick enables optimizing large scale stochastic computation graphs via gradie...
Derivatives, mostly in the form of gradients and Hessians, are ubiquitous in machine learning. Autom...
In the age of artificial intelligence, the best approach to handling huge amounts of data is a treme...
The goal of this paper is to debunk and dispel the magic behind black-box optimizers and stochastic ...
Discrete expectations arise in various machine learning tasks, and we often need to backpropagate th...
Automatic differentiation, as implemented today, does not have a simple mathematical model adapted t...
The score function estimator is widely used for estimating gradients of stochastic objectives in sto...
The deep learning community has devised a diverse set of methods to make gradient optimization, usin...
We introduce the stochastic gradient descent algorithm used in the computational network toolkit (CN...
International audienceStochastic approximation (SA) is a classical algorithm that has had since the ...
This thesis presents five contributions to machine learning, with themes of differentiability and Ba...