Abstract The efficient mapping of program parallelism to multi-core processors is highly dependent on the underlying architecture. This paper proposes a portable and automatic compiler-based approach to mapping such parallelism using machine learning. It develops two predictors: a data sensitive and a data insensitive predictor to select the best mapping for parallel programs. They predict the number of threads and the scheduling policy for any given program using a model learnt off-line. By using low-cost profiling runs, they predict the mapping for a new unseen program across multiple input data sets. We evaluate our approach by selecting parallelism mapping configurations for OpenMP programs on two representative but different multi-core...
OpenMP, a directive-based API supports multithreading programming on shared memory systems. Since O...
: Machine learning using large data sets is a computationally intensive process. One technique that ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
The efficient mapping of program parallelism to multi-core processors is highly dependent on the und...
Multi-core processors are now ubiquitous and are widely seen as the most viable means of delivering ...
Compiler-based auto-parallelization is a much-studied area but has yet to find widespread applicatio...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Single core designs and architectures have reached their limits due to heat and power walls. In orde...
Stream based languages are a popular approach to expressing parallelism in modern applications. The ...
Abstract: Mapping parallel applications to multi-processor architectures requires in-formation about...
Heterogeneous computing systems with multiple CPUs and GPUs are increasingly popular. Today, heterog...
Abstract—Recently parallel architectures have entered every area of computing, from multi-core proce...
New approaches are necessary to generate performance models in current systems due the het erogeneit...
OpenMP, a directive-based API supports multithreading programming on shared memory systems. Since O...
: Machine learning using large data sets is a computationally intensive process. One technique that ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
The efficient mapping of program parallelism to multi-core processors is highly dependent on the und...
Multi-core processors are now ubiquitous and are widely seen as the most viable means of delivering ...
Compiler-based auto-parallelization is a much-studied area but has yet to find widespread applicatio...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Single core designs and architectures have reached their limits due to heat and power walls. In orde...
Stream based languages are a popular approach to expressing parallelism in modern applications. The ...
Abstract: Mapping parallel applications to multi-processor architectures requires in-formation about...
Heterogeneous computing systems with multiple CPUs and GPUs are increasingly popular. Today, heterog...
Abstract—Recently parallel architectures have entered every area of computing, from multi-core proce...
New approaches are necessary to generate performance models in current systems due the het erogeneit...
OpenMP, a directive-based API supports multithreading programming on shared memory systems. Since O...
: Machine learning using large data sets is a computationally intensive process. One technique that ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...