The present invention concerns computer-implemented methods for training a machine learning model using Stochastic Gradient Descent, SGD. In one embodiment, the method is performed by a first computer in a distributed computing environment and comprises performing a learning round, comprising broadcasting a parameter vector to a plurality of worker computers in the distributed computing environment, and upon receipt of one or more respective estimate vectors from a subset of the worker computers, determining an updated parameter vector for use in a next learning round based on the one or more received estimate vectors, wherein the determining comprises ignoring an estimate vector received from a given worker computer when a sending frequenc...
Byzantine-resilient Stochastic Gradient Descent (SGD) aims at shielding model training from Byzantin...
The aim of this thesis is solving minimization problems where the objective function is a sum of a d...
Stochastic Gradient Descent (SGD) is the standard numerical method used to solve the core optimizati...
Asynchronous distributed machine learning solutions have proven very effective so far, but always as...
This paper studies the problem of distributed stochastic optimization in an adversarial setting wher...
This paper considers the Byzantine fault-tolerance problem in distributed stochastic gradient descen...
A very common optimization technique in Machine Learning is Stochastic Gradient Descent (SGD). SGD c...
For many data-intensive real-world applications, such as recognizing objects from images, detecting ...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
We report on \emph{Krum}, the first \emph{provably} Byzantine-tolerant aggregation rule for distribu...
In this paper, we propose a class of robust stochastic subgradient methods for distributed learning ...
We introduce the stochastic gradient descent algorithm used in the computational network toolkit (CN...
We develop a Distributed Event-Triggered Stochastic GRAdient Descent (DETSGRAD) algorithm for solvin...
Distributed machine learning is primarily motivated by the promise of increased computation power fo...
Byzantine-resilient Stochastic Gradient Descent (SGD) aims at shielding model training from Byzantin...
The aim of this thesis is solving minimization problems where the objective function is a sum of a d...
Stochastic Gradient Descent (SGD) is the standard numerical method used to solve the core optimizati...
Asynchronous distributed machine learning solutions have proven very effective so far, but always as...
This paper studies the problem of distributed stochastic optimization in an adversarial setting wher...
This paper considers the Byzantine fault-tolerance problem in distributed stochastic gradient descen...
A very common optimization technique in Machine Learning is Stochastic Gradient Descent (SGD). SGD c...
For many data-intensive real-world applications, such as recognizing objects from images, detecting ...
The implementation of a vast majority of machine learning (ML) algorithms boils down to solving a nu...
Optimization has been the workhorse of solving machine learning problems. However, the efficiency of...
We report on \emph{Krum}, the first \emph{provably} Byzantine-tolerant aggregation rule for distribu...
In this paper, we propose a class of robust stochastic subgradient methods for distributed learning ...
We introduce the stochastic gradient descent algorithm used in the computational network toolkit (CN...
We develop a Distributed Event-Triggered Stochastic GRAdient Descent (DETSGRAD) algorithm for solvin...
Distributed machine learning is primarily motivated by the promise of increased computation power fo...
Byzantine-resilient Stochastic Gradient Descent (SGD) aims at shielding model training from Byzantin...
The aim of this thesis is solving minimization problems where the objective function is a sum of a d...
Stochastic Gradient Descent (SGD) is the standard numerical method used to solve the core optimizati...