Existing dynamic self-scheduling algorithms, used to schedule independent tasks on heterogeneous clusters, cannot handle tasks with dependencies because they lack the support for internode communication. To compensate for this deficiency we introduce a synchronization mechanism that provides inter-processor communication, thus, enabling self-scheduling algorithms to handle efficiently nested loops with dependencies. We also present a weighting mechanism that significantly improves the performance of dynamic self-scheduling algorithms. These algorithms divide the total number of tasks into chunks and assign them to processors. The weighting mechanism adapts the chunk sizes to the computing power and current run-queue state of the processors....
Abstract Loop partitioning on parallel and distributed systems has been a critical problem. Furtherm...
In light of continued advances in loop scheduling, this work revisits the OpenMP loop scheduling by ...
Computationally-intensive loops are the primary source of parallelism in scientific applications. Su...
Efficiently scheduling parallel tasks on to the processors of a shared-memory multiprocessor is crit...
In this paper, we present a new practical processor self-scheduling scheme, Trapezoid Self-Schedulin...
Part 4: Applications of Parallel and Distributed ComputingInternational audienceOrdinary programs co...
Scientific applications often contain large and computationally- intensive parallel loops. Dynamic l...
Distributed Computing Systems are a viable and less ex-pensive alternative to parallel computers. Ho...
[[abstract]]Loop partitioning on parallel and distributed systems has been a critical problem. Furth...
The limitation of vector supercomputing and of device speed has led to the development of multiproce...
Abstract: In this work we present the analysis, on a dynamic processor allocation environment, of fo...
[[abstract]]Recently, more and more studies investigated the is-sue of dealing with the heterogeneit...
Conference of 14th Annual International Conference on Computational Science, ICCS 2014 ; Conference ...
Part 1: Algorithms, Scheduling, Analysis, and Data MiningInternational audienceLoop scheduling schem...
The article of record as published may be found at https://doi.org/10.1007/BF02577870In this paper w...
Abstract Loop partitioning on parallel and distributed systems has been a critical problem. Furtherm...
In light of continued advances in loop scheduling, this work revisits the OpenMP loop scheduling by ...
Computationally-intensive loops are the primary source of parallelism in scientific applications. Su...
Efficiently scheduling parallel tasks on to the processors of a shared-memory multiprocessor is crit...
In this paper, we present a new practical processor self-scheduling scheme, Trapezoid Self-Schedulin...
Part 4: Applications of Parallel and Distributed ComputingInternational audienceOrdinary programs co...
Scientific applications often contain large and computationally- intensive parallel loops. Dynamic l...
Distributed Computing Systems are a viable and less ex-pensive alternative to parallel computers. Ho...
[[abstract]]Loop partitioning on parallel and distributed systems has been a critical problem. Furth...
The limitation of vector supercomputing and of device speed has led to the development of multiproce...
Abstract: In this work we present the analysis, on a dynamic processor allocation environment, of fo...
[[abstract]]Recently, more and more studies investigated the is-sue of dealing with the heterogeneit...
Conference of 14th Annual International Conference on Computational Science, ICCS 2014 ; Conference ...
Part 1: Algorithms, Scheduling, Analysis, and Data MiningInternational audienceLoop scheduling schem...
The article of record as published may be found at https://doi.org/10.1007/BF02577870In this paper w...
Abstract Loop partitioning on parallel and distributed systems has been a critical problem. Furtherm...
In light of continued advances in loop scheduling, this work revisits the OpenMP loop scheduling by ...
Computationally-intensive loops are the primary source of parallelism in scientific applications. Su...