We consider the problem of communication efficient distributed optimization where multiple nodes exchange important algorithm information in every iteration to solve large problems. In particular, we focus on the stochastic variance-reduced gradient and propose a novel approach to make it communication-efficient. That is, we compress the communicated information to a few bits while preserving the linear convergence rate of the original uncompressed algorithm. Comprehensive theoretical and numerical analyses on real datasets reveal that our algorithm can significantly reduce the communication complexity, by as much as 95\%, with almost no noticeable penalty. Moreover, it is much more robust to quantization (in terms of maintaining the true m...
In distributed optimization and machine learning, multiple nodes coordinate to solve large problems....
This dissertation deals with developing optimization algorithms which can be distributed over a netw...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...
We consider the problem of communication efficient distributed optimization where multiple nodes exc...
We consider distributed optimization over several devices, each sending incremental model updates to...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
International audienceWe consider a distributed stochastic optimization problem in networks with fin...
We consider decentralized stochastic optimization with the objective function (e.g. data samples for...
International audienceWe propose distributed algorithms for high-dimensional sparse optimization. In...
With the recent proliferation of large-scale learning problems, there have been a lot of interest o...
In this paper, a novel communication-efficient distributed stochastic algorithm (referred as CO-DSA)...
We establish the O(1/k) convergence rate for distributed stochastic gradient methods that operate ov...
International audienceWe consider the problem of distributed stochastic optimization in networks. Ea...
We develop a Distributed Event-Triggered Stochastic GRAdient Descent (DETSGRAD) algorithm for solvin...
International audienceWe study distributed stochastic gradient (D-SG) method and its accelerated var...
In distributed optimization and machine learning, multiple nodes coordinate to solve large problems....
This dissertation deals with developing optimization algorithms which can be distributed over a netw...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...
We consider the problem of communication efficient distributed optimization where multiple nodes exc...
We consider distributed optimization over several devices, each sending incremental model updates to...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
International audienceWe consider a distributed stochastic optimization problem in networks with fin...
We consider decentralized stochastic optimization with the objective function (e.g. data samples for...
International audienceWe propose distributed algorithms for high-dimensional sparse optimization. In...
With the recent proliferation of large-scale learning problems, there have been a lot of interest o...
In this paper, a novel communication-efficient distributed stochastic algorithm (referred as CO-DSA)...
We establish the O(1/k) convergence rate for distributed stochastic gradient methods that operate ov...
International audienceWe consider the problem of distributed stochastic optimization in networks. Ea...
We develop a Distributed Event-Triggered Stochastic GRAdient Descent (DETSGRAD) algorithm for solvin...
International audienceWe study distributed stochastic gradient (D-SG) method and its accelerated var...
In distributed optimization and machine learning, multiple nodes coordinate to solve large problems....
This dissertation deals with developing optimization algorithms which can be distributed over a netw...
In distributed training of deep models, the transmission volume of stochastic gradients (SG) imposes...