In order to solve a problem in parallel we need to undertake the fundamental step of splitting the computational tasks into parts, i.e. decomposing the problem solving. A whatever decomposition does not necessarily lead to a parallel algorithm with the highest performance. This topic is even more important when complex parallel algorithms must be developed for hybrid or heterogeneous architectures. We present an innovative approach which starts from a problem decomposition into parts (sub-problems). These parts will be regarded as elements of an algebraic structure and will be related to each other according to a suitably defined dependency relationship. The main outcome of such framework is to define a set of block matrices (dependency, de...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
This paper demonstrates that it is possible to obtain good, scalable parallel performance by coordin...
In order to solve a problem in parallel we need to undertake the fundamental step of splitting the c...
(parallel computers and algorithms too). In this sense the paper is devoted to a complex performance...
Processors have become faster at a much quicker rate than memory access time, creating wide gap betw...
AbstractThis paper outlines a theory of parallel algorithms that emphasizes two crucial aspects of p...
In this paper, we describe a model for determining the optimal data and computation decomposition fo...
Performance analysis tools are essential to the maintenance of efficient parallel execution of scien...
With the quickly evolving hardware landscape of high-performance computing (HPC) and its increasing ...
This paper demonstrates that it is possible to obtain good, scalable parallel performance by coordi...
This session explores, through the use of formal methods, the “intuition” used in creating a paralle...
Abstra t. We show in this paper how to evaluate the performan e of pipeline-stru tured parallel prog...
Parallel programming and data-parallel algorithms have been the main techniques supporting high-perf...
Two issues in linear algebra algorithms for multicomputers are addressed. First, how tounify paralle...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
This paper demonstrates that it is possible to obtain good, scalable parallel performance by coordin...
In order to solve a problem in parallel we need to undertake the fundamental step of splitting the c...
(parallel computers and algorithms too). In this sense the paper is devoted to a complex performance...
Processors have become faster at a much quicker rate than memory access time, creating wide gap betw...
AbstractThis paper outlines a theory of parallel algorithms that emphasizes two crucial aspects of p...
In this paper, we describe a model for determining the optimal data and computation decomposition fo...
Performance analysis tools are essential to the maintenance of efficient parallel execution of scien...
With the quickly evolving hardware landscape of high-performance computing (HPC) and its increasing ...
This paper demonstrates that it is possible to obtain good, scalable parallel performance by coordi...
This session explores, through the use of formal methods, the “intuition” used in creating a paralle...
Abstra t. We show in this paper how to evaluate the performan e of pipeline-stru tured parallel prog...
Parallel programming and data-parallel algorithms have been the main techniques supporting high-perf...
Two issues in linear algebra algorithms for multicomputers are addressed. First, how tounify paralle...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
Many physical processes and phenomena in view of their complexity cannot be described analytically. ...
This paper demonstrates that it is possible to obtain good, scalable parallel performance by coordin...