Abstract In the prospect of the upcoming exa-scale era with millions of execution units, the question of how to deal with this level of parallelism efficiently is of time-critical relevance. State-of-the-Art parallelization techniques such as OpenMP and MPI are not guaran-teed to solve the expected problems of starvation, grow-ing latencies, overheads, and contention. On the other hand, new parallelization paradigms promise to effi-ciently hide latencies and contain starvation and con-tention. In this paper we analyze the performance of one novel parallelization strategy for shared and distributed memory machines. We will focus on shared memory ar-chitectures and compare the performance of the Par-alleX execution model against the quasi-sta...
This work examines performance characteristics of multiple shared-memory implementations of a probab...
Ever-increasing core counts create the need to develop parallel algorithms that avoid closely- coupl...
Nowadays, shared-memory parallel architectures have evolved and new programming frameworks have appe...
Machines comprised of a distributed collection of shared memory or SMP nodes are becoming common for...
High performance computing (HPC) is experiencing a phase change with the challenges of programming a...
High Performance Computing (HPC) can be defined as the practice of combining computing power to atta...
Task and/or data parallelism can be exploited in most applications on today's multi-core processors....
Communicated by Guest Editors The implementation of stencil computations on modern, massively parall...
The majority of current HPC applications are composed of complex and irregular data structures that ...
This paper proposes the study of a new computation model that attempts to address the underlying sou...
In prior work, we have proposed techniques to extend the ease of shared-memory parallel programming ...
The paper introduces an OpenMP implementation of pipelined Parareal and compares it to a standard MP...
In prior work, we have proposed techniques to extend the ease of shared-memory parallel programming ...
Abstract. In this paper, we present ParaPART, a parallel version of a mesh partitioning tool, called...
OpenMP has established itself as the de facto standard for parallel programming on shared-memory pla...
This work examines performance characteristics of multiple shared-memory implementations of a probab...
Ever-increasing core counts create the need to develop parallel algorithms that avoid closely- coupl...
Nowadays, shared-memory parallel architectures have evolved and new programming frameworks have appe...
Machines comprised of a distributed collection of shared memory or SMP nodes are becoming common for...
High performance computing (HPC) is experiencing a phase change with the challenges of programming a...
High Performance Computing (HPC) can be defined as the practice of combining computing power to atta...
Task and/or data parallelism can be exploited in most applications on today's multi-core processors....
Communicated by Guest Editors The implementation of stencil computations on modern, massively parall...
The majority of current HPC applications are composed of complex and irregular data structures that ...
This paper proposes the study of a new computation model that attempts to address the underlying sou...
In prior work, we have proposed techniques to extend the ease of shared-memory parallel programming ...
The paper introduces an OpenMP implementation of pipelined Parareal and compares it to a standard MP...
In prior work, we have proposed techniques to extend the ease of shared-memory parallel programming ...
Abstract. In this paper, we present ParaPART, a parallel version of a mesh partitioning tool, called...
OpenMP has established itself as the de facto standard for parallel programming on shared-memory pla...
This work examines performance characteristics of multiple shared-memory implementations of a probab...
Ever-increasing core counts create the need to develop parallel algorithms that avoid closely- coupl...
Nowadays, shared-memory parallel architectures have evolved and new programming frameworks have appe...