The Map-Reduce computing framework rose to prominence with datasets of such size that dozens of machines on a single cluster were needed for individual jobs. As datasets approach the exabyte scale, a single job may need distributed processing not only on multiple machines, but on multiple clusters. We consider a scheduling problem to minimize weighted average completion time of n jobs on m distributed clusters of parallel machines. In keeping with the scale of the problems motivating this work, we assume that (1) each job is divided into m "subjobs" and (2) distinct subjobs of a given job may be processed concurrently. When each cluster is a single machine, this is the NP-Hard concurrent open shop problem. A clear limitation of such a mod...
EuroPar 2012In this paper we tackle the well-known problem of scheduling a collection of parallel jo...
Scheduling is a crucial problem in parallel and distributed processing. It consists of determining w...
We consider a scheduling problem where a set of jobs is a-priori distributed over parallel machines....
The Map-Reduce computing framework rose to prominence with datasets of such size that dozens of mach...
International audienceThe Multiple Cluster Scheduling Problem corresponds to minimize the maximum co...
International audienceWe consider the Multiple Cluster Scheduling Problem (MCSP), where the objectiv...
International audienceMany scientific applications can be structured as Parallel Task Graphs (PTGs),...
MapReduce framework is established as the standard approach for parallel processing of massive amoun...
In malleable job scheduling, jobs can be executed simultaneously on multiple machines with the proce...
Recent years have witnessed a rapid growth of distributed machine learning (ML) frameworks, which ex...
International audienceIn this paper, we tackle the well‐known problem of scheduling a collection of ...
© 2017 Springer Science+Business Media New York We consider the problem of scheduling a number of jo...
We study the problem of scheduling jobs on parallel machines minimizing the total completion time, w...
International audienceApplications structured as parallel task graphs exhibit both data and task par...
International audienceMany scientific applications can be structured as Parallel Task Graphs (PTGs),...
EuroPar 2012In this paper we tackle the well-known problem of scheduling a collection of parallel jo...
Scheduling is a crucial problem in parallel and distributed processing. It consists of determining w...
We consider a scheduling problem where a set of jobs is a-priori distributed over parallel machines....
The Map-Reduce computing framework rose to prominence with datasets of such size that dozens of mach...
International audienceThe Multiple Cluster Scheduling Problem corresponds to minimize the maximum co...
International audienceWe consider the Multiple Cluster Scheduling Problem (MCSP), where the objectiv...
International audienceMany scientific applications can be structured as Parallel Task Graphs (PTGs),...
MapReduce framework is established as the standard approach for parallel processing of massive amoun...
In malleable job scheduling, jobs can be executed simultaneously on multiple machines with the proce...
Recent years have witnessed a rapid growth of distributed machine learning (ML) frameworks, which ex...
International audienceIn this paper, we tackle the well‐known problem of scheduling a collection of ...
© 2017 Springer Science+Business Media New York We consider the problem of scheduling a number of jo...
We study the problem of scheduling jobs on parallel machines minimizing the total completion time, w...
International audienceApplications structured as parallel task graphs exhibit both data and task par...
International audienceMany scientific applications can be structured as Parallel Task Graphs (PTGs),...
EuroPar 2012In this paper we tackle the well-known problem of scheduling a collection of parallel jo...
Scheduling is a crucial problem in parallel and distributed processing. It consists of determining w...
We consider a scheduling problem where a set of jobs is a-priori distributed over parallel machines....