The reparameterization trick enables optimizing large scale stochastic computation graphs via gradient descent. The essence of the trick is to refactor each stochastic node into a differentiable function of its parameters and a random variable with fixed distribution. After refactoring, the gradients of the loss propagated by the chain rule through the graph are low variance unbiased estimators of the gradients of the expected loss. While many continuous random variables have such reparameterizations, discrete random variables lack useful reparameterizations due to the discontinuous nature of discrete states. In this work we introduce CONCRETE random variables—CONtinuous relaxations of disCRETE random variables. The Concrete distribution is...
The direct application of statistics to stochastic optimization based on iterated density estimation...
International audienceWe study the optimization of a continuous function by its stochastic relaxatio...
A central problem in statistical learning is to design prediction algorithms that not only perform w...
The reparameterization trick enables optimizing large scale stochastic computation graphs via gradie...
Numerous models for supervised and reinforcement learning benefit from combinations of discrete and ...
We examine properties of the Concrete (or Gumbel-softmax) distribution on the simplex. Using the nat...
Modelers use automatic differentiation (AD) of computation graphs to implement complex deep learning...
Continuous relaxations play an important role in discrete optimization, but have not seen much use i...
By enabling correct differentiation in Stochastic Computation Graphs (SCGs), the infinitely differen...
The paper studies a stochastic extension of continuous recurrent neural networks and analyzes gradie...
The paper studies a stochastic extension of continuous recurrent neural networks and analyzes gradie...
Discrete expectations arise in various machine learning tasks, and we often need to backpropagate th...
We present a tree-based reparameterization framework for the approximate estimation of stochastic pr...
Slides of a talk given at Dortmund University, Dept. of Statistics, on March 2015 the 11th. Invitati...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
The direct application of statistics to stochastic optimization based on iterated density estimation...
International audienceWe study the optimization of a continuous function by its stochastic relaxatio...
A central problem in statistical learning is to design prediction algorithms that not only perform w...
The reparameterization trick enables optimizing large scale stochastic computation graphs via gradie...
Numerous models for supervised and reinforcement learning benefit from combinations of discrete and ...
We examine properties of the Concrete (or Gumbel-softmax) distribution on the simplex. Using the nat...
Modelers use automatic differentiation (AD) of computation graphs to implement complex deep learning...
Continuous relaxations play an important role in discrete optimization, but have not seen much use i...
By enabling correct differentiation in Stochastic Computation Graphs (SCGs), the infinitely differen...
The paper studies a stochastic extension of continuous recurrent neural networks and analyzes gradie...
The paper studies a stochastic extension of continuous recurrent neural networks and analyzes gradie...
Discrete expectations arise in various machine learning tasks, and we often need to backpropagate th...
We present a tree-based reparameterization framework for the approximate estimation of stochastic pr...
Slides of a talk given at Dortmund University, Dept. of Statistics, on March 2015 the 11th. Invitati...
Many connectionist learning algorithms consists of minimizing a cost of the form C(w) = E(J(z; w)) ...
The direct application of statistics to stochastic optimization based on iterated density estimation...
International audienceWe study the optimization of a continuous function by its stochastic relaxatio...
A central problem in statistical learning is to design prediction algorithms that not only perform w...