In distributed computing, slower nodes (stragglers) usually become a bottleneck. Gradient Coding (GC), introduced by Tandon et al., is an efficient technique that uses principles of error-correcting codes to distribute gradient computation in the presence of stragglers. In this paper, we consider the distributed computation of a sequence of gradients $\{g(1),g(2),\ldots,g(J)\}$, where processing of each gradient $g(t)$ starts in round-$t$ and finishes by round-$(t+T)$. Here $T\geq 0$ denotes a delay parameter. For the GC scheme, coding is only across computing nodes and this results in a solution where $T=0$. On the other hand, having $T>0$ allows for designing schemes which exploit the temporal dimension as well. In this work, we propose t...
This dissertation considers distributed algorithms for centralized and decentralized networks that s...
We present two novel federated learning (FL) schemes that mitigate the effect of straggling devices ...
Parallel implementations of stochastic gradient descent (SGD) have received significant research att...
Distributed implementations are crucial in speeding up large scale machine learning applications. Di...
In distributed synchronous gradient descent (GD) the main performance bottleneck for the per-iterati...
When gradient descent (GD) is scaled to many parallel computing servers (workers) for large scale ma...
Distributed implementations are crucial in speeding up large scale machine learning applications. Di...
When gradient descent (GD) is scaled to many parallel workers for large-scale machine learning appli...
Gradient coding is a technique for straggler mitigation in distributed learning. In this paper we de...
Gradient descent (GD) methods are commonly employed in machine learning problems to optimize the par...
Today's massively-sized datasets have made it necessary to often perform computations on them in a d...
We consider the setting where a master wants to run a distributed stochastic gradient descent (SGD) ...
As the size of models and datasets grows, it has become increasingly common to train models in paral...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
Huge scale machine learning problems are nowadays tackled by distributed optimization algorithms, i....
This dissertation considers distributed algorithms for centralized and decentralized networks that s...
We present two novel federated learning (FL) schemes that mitigate the effect of straggling devices ...
Parallel implementations of stochastic gradient descent (SGD) have received significant research att...
Distributed implementations are crucial in speeding up large scale machine learning applications. Di...
In distributed synchronous gradient descent (GD) the main performance bottleneck for the per-iterati...
When gradient descent (GD) is scaled to many parallel computing servers (workers) for large scale ma...
Distributed implementations are crucial in speeding up large scale machine learning applications. Di...
When gradient descent (GD) is scaled to many parallel workers for large-scale machine learning appli...
Gradient coding is a technique for straggler mitigation in distributed learning. In this paper we de...
Gradient descent (GD) methods are commonly employed in machine learning problems to optimize the par...
Today's massively-sized datasets have made it necessary to often perform computations on them in a d...
We consider the setting where a master wants to run a distributed stochastic gradient descent (SGD) ...
As the size of models and datasets grows, it has become increasingly common to train models in paral...
We consider the distributed stochastic gradient descent problem, where a main node distributes gradi...
Huge scale machine learning problems are nowadays tackled by distributed optimization algorithms, i....
This dissertation considers distributed algorithms for centralized and decentralized networks that s...
We present two novel federated learning (FL) schemes that mitigate the effect of straggling devices ...
Parallel implementations of stochastic gradient descent (SGD) have received significant research att...