This paper studies the problem of distributed stochastic optimization in an adversarial setting where, out of m machines which allegedly compute stochastic gradients every iteration, an α-fraction are Byzantine, and may behave adversarially. Our main result is a variant of stochastic gradient descent (SGD) which finds ε-approximate minimizers of convex functions in T=O~(1/ε²m+α²/ε²) iterations. In contrast, traditional mini-batch SGD needs T=O(1/ε²m) iterations, but cannot tolerate Byzantine failures. Further, we provide a lower bound showing that, up to logarithmic factors, our algorithm is information-theoretically optimal both in terms of sample complexity and time complexity
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
A very common optimization technique in Machine Learning is Stochastic Gradient Descent (SGD). SGD c...
Stochastic and data-distributed optimization algorithms have received lots of attention from the mac...
For many data-intensive real-world applications, such as recognizing objects from images, detecting ...
Asynchronous distributed machine learning solutions have proven very effective so far, but always as...
Byzantine-resilient Stochastic Gradient Descent (SGD) aims at shielding model training from Byzantin...
This paper considers the Byzantine fault-tolerance problem in distributed stochastic gradient descen...
Abstract. In this paper we study the effect of stochastic errors on two constrained incremental sub-...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
While machine learning is going through an era of celebrated success, concerns have been raised abou...
The present invention concerns computer-implemented methods for training a machine learning model us...
We report on \emph{Krum}, the first \emph{provably} Byzantine-tolerant aggregation rule for distribu...
This work focuses on decentralized stochastic optimization in the presence of Byzantine attacks. Dur...
International audienceWe study distributed stochastic gradient (D-SG) method and its accelerated var...
We establish the O(1/k) convergence rate for distributed stochastic gradient methods that operate ov...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
A very common optimization technique in Machine Learning is Stochastic Gradient Descent (SGD). SGD c...
Stochastic and data-distributed optimization algorithms have received lots of attention from the mac...
For many data-intensive real-world applications, such as recognizing objects from images, detecting ...
Asynchronous distributed machine learning solutions have proven very effective so far, but always as...
Byzantine-resilient Stochastic Gradient Descent (SGD) aims at shielding model training from Byzantin...
This paper considers the Byzantine fault-tolerance problem in distributed stochastic gradient descen...
Abstract. In this paper we study the effect of stochastic errors on two constrained incremental sub-...
Stochastic gradient descent (SGD) is a simple and popular method to solve stochastic optimization pr...
While machine learning is going through an era of celebrated success, concerns have been raised abou...
The present invention concerns computer-implemented methods for training a machine learning model us...
We report on \emph{Krum}, the first \emph{provably} Byzantine-tolerant aggregation rule for distribu...
This work focuses on decentralized stochastic optimization in the presence of Byzantine attacks. Dur...
International audienceWe study distributed stochastic gradient (D-SG) method and its accelerated var...
We establish the O(1/k) convergence rate for distributed stochastic gradient methods that operate ov...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
A very common optimization technique in Machine Learning is Stochastic Gradient Descent (SGD). SGD c...
Stochastic and data-distributed optimization algorithms have received lots of attention from the mac...