Nested parallelism is a well-known parallelization strategy to exploit irregular parallelism in HPC applications. This strategy also fits in critical real-time embedded systems, composed of a set of concurrent functionalities. In this case, nested parallelism can be used to further exploit the parallelism of each functionality. However, current run-time implementations of nested parallelism can produce inefficiencies and load imbalance. Moreover, in critical real-time embedded systems, it may lead to incorrect executions due to, for instance, a work non-conserving scheduler. In both cases, the reason is that the teams of OpenMP threads are a black-box for the scheduler, i.e., the scheduler that assigns OpenMP threads and tasks to the set of...
The development of networksand multi-processor computers has allowed us to solve problems in paralle...
A chief characteristic of next-generation computing systems is the prevalence of parallelism at mult...
Today, almost all desktop and laptop computers are shared-memory multicores, but the code they run i...
Nested parallelism is a well-known parallelization strategy to exploit irregular parallelism in HPC ...
Emerging architecture designs include tens of processing cores on a single chip die; it is believed ...
The recent addition of task parallelism to the OpenMP shared memory API allows programmers to expres...
Due to power constraints, future growth in computing capability must explicitly leverage parallelism...
Modern designs for embedded systems are increasingly embracing cluster-based architectures, where sm...
Tasking promises a model to program parallel applications that provides intuitive semantics. In the ...
In this paper we discuss the use of nested parallelism. Our claim is that if the problem naturally p...
Choosing the appropriate assignment of loop iterations to threads is one of the most important decis...
We develop real-time scheduling techniques for improving performance and energy for multiprogrammed ...
A chief characteristic of next-generation computing systems is the prevalence of parallelism at mult...
CPS Student Forum Portugal was held as part of the Cyber-Physical Systems Week (CPS Week 2018), 10-1...
Chip-level multiprocessors (CMP) have multiple processing cores (Cores) and generally have their cac...
The development of networksand multi-processor computers has allowed us to solve problems in paralle...
A chief characteristic of next-generation computing systems is the prevalence of parallelism at mult...
Today, almost all desktop and laptop computers are shared-memory multicores, but the code they run i...
Nested parallelism is a well-known parallelization strategy to exploit irregular parallelism in HPC ...
Emerging architecture designs include tens of processing cores on a single chip die; it is believed ...
The recent addition of task parallelism to the OpenMP shared memory API allows programmers to expres...
Due to power constraints, future growth in computing capability must explicitly leverage parallelism...
Modern designs for embedded systems are increasingly embracing cluster-based architectures, where sm...
Tasking promises a model to program parallel applications that provides intuitive semantics. In the ...
In this paper we discuss the use of nested parallelism. Our claim is that if the problem naturally p...
Choosing the appropriate assignment of loop iterations to threads is one of the most important decis...
We develop real-time scheduling techniques for improving performance and energy for multiprogrammed ...
A chief characteristic of next-generation computing systems is the prevalence of parallelism at mult...
CPS Student Forum Portugal was held as part of the Cyber-Physical Systems Week (CPS Week 2018), 10-1...
Chip-level multiprocessors (CMP) have multiple processing cores (Cores) and generally have their cac...
The development of networksand multi-processor computers has allowed us to solve problems in paralle...
A chief characteristic of next-generation computing systems is the prevalence of parallelism at mult...
Today, almost all desktop and laptop computers are shared-memory multicores, but the code they run i...