In this paper, we show the potential benefits of translating OpenMP code to low-level parallel code using a data flow execution model, instead of targeting it directly to a multi-threaded program. Our goal is to improve data locality as well as reduce synchronization overheads without introducing data distribution directives to OpenMP. We outline an API that enables us to realize this model using SMARTS (Shared Memory Asynchronous Run-Time System), describe the work of the compiler and discuss the benefits of translating OpenMP to parallel code using data flow execution model. We show experimental results based part of the Parallel Ocean Program (POP) code and Jacobi kernel code running on an SGI Origin 2000. 1
In this paper we will make an experimental description of the parallel programming using OpenMP. Usi...
Selected for presentation at the HiPEAC 2013 Conf.International audienceWe present OpenStream, a dat...
OpenMP has emerged as the de facto standard for writing parallel programs on shared address space pl...
In this paper, we present an alternative implementation of the NANOS OpenMP runtime library (NthLib)...
OpenMP has emerged as an important model and language extension for shared-memory parallel programmi...
Shared memory parallel programming, for instance by inserting OpenMP pragmas into program code, migh...
We have developed compiler optimization techniques for explicit parallel programs using the OpenMP A...
OpenMP has established itself as the de facto standard for parallel programming on shared-memory pla...
OpenMP is attracting wide-spread interest because of its easy-to-use parallel programming model for ...
OpenMP is a very convenient programming model for critical real-time parallel applications due to it...
We present OpenStream, a data-flow extension of OpenMP to express dynamic dependent tasks. The lan-g...
Abstract. This paper presents a source-to-source translation strategy from OpenMP to Global Arrays i...
Locality of computation is key to obtaining high performance on a broad variety of parallel architec...
International audienceTo efficiently exploit high performance computing platforms, applications curr...
The OpenMP Application Programming Interface (API) is an emerging standard for parallel programming ...
In this paper we will make an experimental description of the parallel programming using OpenMP. Usi...
Selected for presentation at the HiPEAC 2013 Conf.International audienceWe present OpenStream, a dat...
OpenMP has emerged as the de facto standard for writing parallel programs on shared address space pl...
In this paper, we present an alternative implementation of the NANOS OpenMP runtime library (NthLib)...
OpenMP has emerged as an important model and language extension for shared-memory parallel programmi...
Shared memory parallel programming, for instance by inserting OpenMP pragmas into program code, migh...
We have developed compiler optimization techniques for explicit parallel programs using the OpenMP A...
OpenMP has established itself as the de facto standard for parallel programming on shared-memory pla...
OpenMP is attracting wide-spread interest because of its easy-to-use parallel programming model for ...
OpenMP is a very convenient programming model for critical real-time parallel applications due to it...
We present OpenStream, a data-flow extension of OpenMP to express dynamic dependent tasks. The lan-g...
Abstract. This paper presents a source-to-source translation strategy from OpenMP to Global Arrays i...
Locality of computation is key to obtaining high performance on a broad variety of parallel architec...
International audienceTo efficiently exploit high performance computing platforms, applications curr...
The OpenMP Application Programming Interface (API) is an emerging standard for parallel programming ...
In this paper we will make an experimental description of the parallel programming using OpenMP. Usi...
Selected for presentation at the HiPEAC 2013 Conf.International audienceWe present OpenStream, a dat...
OpenMP has emerged as the de facto standard for writing parallel programs on shared address space pl...