Stochastic mirror descent (SMD) algorithms have recently garnered a great deal of attention in optimization, signal processing, and machine learning. They are similar to stochastic gradient descent (SGD), in that they perform updates along the negative gradient of an instantaneous (or stochastically chosen) loss function. However, rather than update the parameter (or weight) vector directly, they update it in a "mirrored" domain whose transformation is given by the gradient of a strictly convex differentiable potential function. SMD was originally conceived to take advantage of the underlying geometry of the problem as a way to improve the convergence rate over SGD. In this paper, we study SMD, for linear models and convex loss functions, t...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...
Stochastic mirror descent (SMD) algorithms have recently garnered a great deal of attention in optim...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
Stochastic mirror descent (SMD) is a fairly new family of algorithms that has recently found a wide ...
Stochastic mirror descent (SMD) is a fairly new family of algorithms that has recently found a wide ...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
The stochastic mirror descent (SMD) algorithm is a general class of training algorithms, which inclu...
International audienceIn this paper, we examine a class of nonconvex stochastic opti-mization proble...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
In this paper, we discuss application of iterative Stochastic Optimization routines to the problem o...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...
Stochastic mirror descent (SMD) algorithms have recently garnered a great deal of attention in optim...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
We study the convergence, the implicit regularization and the generalization of stochastic mirror de...
Stochastic mirror descent (SMD) is a fairly new family of algorithms that has recently found a wide ...
Stochastic mirror descent (SMD) is a fairly new family of algorithms that has recently found a wide ...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
Stochastic descent methods (of the gradient and mirror varieties) have become increasingly popular i...
The stochastic mirror descent (SMD) algorithm is a general class of training algorithms, which inclu...
International audienceIn this paper, we examine a class of nonconvex stochastic opti-mization proble...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
In this paper, we discuss application of iterative Stochastic Optimization routines to the problem o...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
International audienceIn this paper, we examine a class of non-convex stochastic optimization proble...