This paper considers an analytic data distribution for improving the performance of host-client type parallel applications which exhibit serialized communication patterns. The technique involves assuming the serialized communications is enforced, which simplifies data analysis and can provide the basis for real-time dynamic load balancing. This distribution has been tested using a parallel matrix multiplication implementation and a parallel MPEG compression implementation. The key results of this paper are that analytic distribution can reduce execution time and increase scalability of certain parallel applications over typical equal data distributions
Parallelizing sparse irregular application on distributed memory systems poses serious scalability c...
The performance of a High Performance Parallel or Distributed Computation depends heavily on minimiz...
International audienceThis paper discusses some algorithmic issues when computing with a heterogeneo...
This paper presents a reusable design of a data distribution frame-work for data parallel high perfo...
Parallel systems leverage parallel file systems to efficiently perform I/O to shared files. These pa...
Abstract: High performance computing (HPC) architectures are specialized machines which can reach th...
The paper presents the mechanisms for dynamic load distribution implemented within the support for t...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Data distribution is one of the key aspects that a parallelizing compiler for a distributed memory a...
A parallel concurrent application runs most efficiently and quickly when the workload is distributed...
The goal of high performance computing is executing very large problems in the least amount of time,...
We present algorithms for the transportation of data in parallel and distributed systems that would ...
(eng) We describe the compilation and execution of data-parallel languages for networks of workstati...
High performance computing (HPC) architectures are specialized machines which can reach their peak p...
Load distribution is essential for efficient use of available processors in a parallel branch-and-bo...
Parallelizing sparse irregular application on distributed memory systems poses serious scalability c...
The performance of a High Performance Parallel or Distributed Computation depends heavily on minimiz...
International audienceThis paper discusses some algorithmic issues when computing with a heterogeneo...
This paper presents a reusable design of a data distribution frame-work for data parallel high perfo...
Parallel systems leverage parallel file systems to efficiently perform I/O to shared files. These pa...
Abstract: High performance computing (HPC) architectures are specialized machines which can reach th...
The paper presents the mechanisms for dynamic load distribution implemented within the support for t...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Data distribution is one of the key aspects that a parallelizing compiler for a distributed memory a...
A parallel concurrent application runs most efficiently and quickly when the workload is distributed...
The goal of high performance computing is executing very large problems in the least amount of time,...
We present algorithms for the transportation of data in parallel and distributed systems that would ...
(eng) We describe the compilation and execution of data-parallel languages for networks of workstati...
High performance computing (HPC) architectures are specialized machines which can reach their peak p...
Load distribution is essential for efficient use of available processors in a parallel branch-and-bo...
Parallelizing sparse irregular application on distributed memory systems poses serious scalability c...
The performance of a High Performance Parallel or Distributed Computation depends heavily on minimiz...
International audienceThis paper discusses some algorithmic issues when computing with a heterogeneo...