International audienceWe introduce the "continuized" Nesterov acceleration, a close variant of Nesterov acceleration whose variables are indexed by a continuous time parameter. The two variables continuously mix following a linear ordinary differential equation and take gradient steps at random times. This continuized variant benefits from the best of the continuous and the discrete frameworks: as a continuous process, one can use differential calculus to analyze convergence and obtain analytical expressions for the parameters; and a discretization of the continuized process can be computed exactly with convergence rates similar to those of Nesterov original acceleration. We show that the discretization has the same structure as Nesterov ac...
We consider the stochastic approximation problem in a streaming framework where an objective is mini...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
<p>We consider distributed optimization in random networks where N nodes cooperatively minimize the ...
International audienceWe introduce the "continuized" Nesterov acceleration, a close variant of Neste...
We introduce the "continuized" Nesterov acceleration, a close variant of Nesterov acceleration whose...
We develop the mathematical foundations of the stochastic modified equations (SME) framework for ana...
Motivated by the recent interest in statistical learning and distributed computing, we study stochas...
International audienceIn this paper, a general stochastic optimization procedure is studied, unifyin...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Abstract: Stochastic gradient descent is an optimisation method that combines classical gradient des...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
Smoothed functional (SF) algorithm estimates the gradient of the stochastic optimization problem by ...
We consider distributed optimization where N nodes in a generic, connected network minimize the sum ...
In this article a family of second order ODEs associated to inertial gradient descend is studied. Th...
We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of...
We consider the stochastic approximation problem in a streaming framework where an objective is mini...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
<p>We consider distributed optimization in random networks where N nodes cooperatively minimize the ...
International audienceWe introduce the "continuized" Nesterov acceleration, a close variant of Neste...
We introduce the "continuized" Nesterov acceleration, a close variant of Nesterov acceleration whose...
We develop the mathematical foundations of the stochastic modified equations (SME) framework for ana...
Motivated by the recent interest in statistical learning and distributed computing, we study stochas...
International audienceIn this paper, a general stochastic optimization procedure is studied, unifyin...
Current machine learning practice requires solving huge-scale empirical risk minimization problems q...
Abstract: Stochastic gradient descent is an optimisation method that combines classical gradient des...
We design step-size schemes that make stochastic gradient descent (SGD) adaptive to (i) the noise σ ...
Smoothed functional (SF) algorithm estimates the gradient of the stochastic optimization problem by ...
We consider distributed optimization where N nodes in a generic, connected network minimize the sum ...
In this article a family of second order ODEs associated to inertial gradient descend is studied. Th...
We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of...
We consider the stochastic approximation problem in a streaming framework where an objective is mini...
International audienceIn this paper, we propose a unified view of gradient-based algorithms for stoc...
<p>We consider distributed optimization in random networks where N nodes cooperatively minimize the ...