Abstract. Parallelism is one of the main sources for performance improvement in modern computing environment, but the efficient exploitation of the available parallelism depends on a number of parameters. Determining the optimum number of threads for a given data parallel loop, for example, is a difficult problem and dependent on the specific parallel platform. This paper presents a learning-based approach to parallel workload allocation in a costaware manner. This approach uses static program features to classify programs, before deciding the best workload allocation scheme based on its prior experience with similar programs. Experimental results on 12 Java benchmarks (76 test cases with different workloads in total) show that it can effic...
Fast response, storage efficiency, fault tolerance and graceful degradation in face of scarce or spu...
International audienceThe performance of irregular scientific applications can be easily affected by...
The performance of parallel code significantly depends on the parallel task granularity (PTG). If th...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Abstract The efficient mapping of program parallelism to multi-core processors is highly dependent o...
International audienceThis paper presents the parallelization of a machine learning method, called t...
This paper introduces a reinforcement-learning based resource allocation framework for dynamic place...
: Machine learning using large data sets is a computationally intensive process. One technique that ...
This paper introduces a resource allocation framework specifically tailored for addressing the probl...
from object-oriented programming techniques because of their flexible and modular program developmen...
This paper describes a dynamic framework for mapping the threads of parallel applications to the com...
We describe in this paper a new approach to parallelize branch-and-bound on a certain number of proc...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
This paper describes a dynamic framework for mapping the threads of parallel applications to the com...
This paper describes a portable, machine learning-based approach to Java optimisation. This approach...
Fast response, storage efficiency, fault tolerance and graceful degradation in face of scarce or spu...
International audienceThe performance of irregular scientific applications can be easily affected by...
The performance of parallel code significantly depends on the parallel task granularity (PTG). If th...
Modern day hardware platforms are parallel and diverse, ranging from mobiles to data centers. Mains...
Abstract The efficient mapping of program parallelism to multi-core processors is highly dependent o...
International audienceThis paper presents the parallelization of a machine learning method, called t...
This paper introduces a reinforcement-learning based resource allocation framework for dynamic place...
: Machine learning using large data sets is a computationally intensive process. One technique that ...
This paper introduces a resource allocation framework specifically tailored for addressing the probl...
from object-oriented programming techniques because of their flexible and modular program developmen...
This paper describes a dynamic framework for mapping the threads of parallel applications to the com...
We describe in this paper a new approach to parallelize branch-and-bound on a certain number of proc...
Much compiler-orientated work in the area of mapping parallel programs to parallel architectures has...
This paper describes a dynamic framework for mapping the threads of parallel applications to the com...
This paper describes a portable, machine learning-based approach to Java optimisation. This approach...
Fast response, storage efficiency, fault tolerance and graceful degradation in face of scarce or spu...
International audienceThe performance of irregular scientific applications can be easily affected by...
The performance of parallel code significantly depends on the parallel task granularity (PTG). If th...