The efficient mapping of program parallelism to multi-core processors is highly dependent on the underlying architecture. This paper proposes a portable and automatic compiler-based approach to mapping such parallelism using machine learning. It develops two predictors: a data sensitive and a data insensitive predictor to select the best mapping for parallel programs. They predict the number of threads and the scheduling policy for any given program using a model learnt off-line. By using low-cost profiling runs, they predict the mapping for a new unseen program across multiple input data sets. We evaluate our approach by selecting parallelism mapping configurations for OpenMP programs on two representative but different multi-core platform...
New approaches are necessary to generate performance models in current systems due the het erogeneit...
International audienceThe parallelism in shared-memory systems has increased significantly with the ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
Abstract The efficient mapping of program parallelism to multi-core processors is highly dependent o...
Multi-core processors are now ubiquitous and are widely seen as the most viable means of delivering ...
Compiler-based auto-parallelization is a much-studied area but has yet to find widespread applicatio...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Single core designs and architectures have reached their limits due to heat and power walls. In orde...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Stream based languages are a popular approach to expressing parallelism in modern applications. The ...
Heterogeneous computing systems with multiple CPUs and GPUs are increasingly popular. Today, heterog...
Abstract—Recently parallel architectures have entered every area of computing, from multi-core proce...
OpenMP, a directive-based API supports multithreading programming on shared memory systems. Since O...
Abstract: Mapping parallel applications to multi-processor architectures requires in-formation about...
New approaches are necessary to generate performance models in current systems due the het erogeneit...
International audienceThe parallelism in shared-memory systems has increased significantly with the ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
Abstract The efficient mapping of program parallelism to multi-core processors is highly dependent o...
Multi-core processors are now ubiquitous and are widely seen as the most viable means of delivering ...
Compiler-based auto-parallelization is a much-studied area but has yet to find widespread applicatio...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Compiler-based auto-parallelization is a much studied area, yet has still not found wide-spread appl...
Single core designs and architectures have reached their limits due to heat and power walls. In orde...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Stream based languages are a popular approach to expressing parallelism in modern applications. The ...
Heterogeneous computing systems with multiple CPUs and GPUs are increasingly popular. Today, heterog...
Abstract—Recently parallel architectures have entered every area of computing, from multi-core proce...
OpenMP, a directive-based API supports multithreading programming on shared memory systems. Since O...
Abstract: Mapping parallel applications to multi-processor architectures requires in-formation about...
New approaches are necessary to generate performance models in current systems due the het erogeneit...
International audienceThe parallelism in shared-memory systems has increased significantly with the ...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...