The rapidly evolving landscape of multicore architectures makes the construction of efficient libraries a daunting task. A family of methods known collectively as “auto-tuning” has emerged to address this challenge. Two major approaches to auto-tuning are empirical and model-based: empirical autotuning is a generic but slow approach that works by measuring runtimes of candidate implementations, model-based auto-tuning predicts those runtimes using simplified abstractions designed by hand. We show that machine learning methods for non-linear regression can be used to estimate timing models from data, capturing the best of both approaches. A statistically-derived model offers the speed of a model-based approach, with the generality and simpli...
We develop methods for adjusting device configurations to runtime conditions based on system-state p...
Hyperparameter optimization is crucial for achieving peak performance with many machine learning alg...
International audienceAnalyzing better time series with limited human effort is of interest to acade...
The tuning of learning algorithm parameters has become more and more important during the last years...
Mathematical solvers have evolved to become complex software and thereby have become a difficult sub...
We present a novel strategy for automatic performance tuning of GPU programs. The strategy combines ...
Abstract. Machine learning can be utilized to build models that predict the runtime of search algori...
Auto-Tuning DL compilers are gaining ground as an optimizing back-end for DL frameworks. While exist...
In high-performance computing, excellent node-level performance is required for the efficient use of...
Over the past couple of decades, many telecommunication industries have passed through the different...
Achieving peak performance from library subroutines usually requires extensive, machine-dependent tu...
Our pipeline can be separated into three parts: (i) initial data preparation, (ii) training and pred...
The ability to handle and analyse massive amounts of data has been progressively improved during the...
International audienceAutotuning, the practice of automatic tuning of applications to provide perfor...
Abstract. Multicore architectures featuring specialized accelerators are getting an increasing amoun...
We develop methods for adjusting device configurations to runtime conditions based on system-state p...
Hyperparameter optimization is crucial for achieving peak performance with many machine learning alg...
International audienceAnalyzing better time series with limited human effort is of interest to acade...
The tuning of learning algorithm parameters has become more and more important during the last years...
Mathematical solvers have evolved to become complex software and thereby have become a difficult sub...
We present a novel strategy for automatic performance tuning of GPU programs. The strategy combines ...
Abstract. Machine learning can be utilized to build models that predict the runtime of search algori...
Auto-Tuning DL compilers are gaining ground as an optimizing back-end for DL frameworks. While exist...
In high-performance computing, excellent node-level performance is required for the efficient use of...
Over the past couple of decades, many telecommunication industries have passed through the different...
Achieving peak performance from library subroutines usually requires extensive, machine-dependent tu...
Our pipeline can be separated into three parts: (i) initial data preparation, (ii) training and pred...
The ability to handle and analyse massive amounts of data has been progressively improved during the...
International audienceAutotuning, the practice of automatic tuning of applications to provide perfor...
Abstract. Multicore architectures featuring specialized accelerators are getting an increasing amoun...
We develop methods for adjusting device configurations to runtime conditions based on system-state p...
Hyperparameter optimization is crucial for achieving peak performance with many machine learning alg...
International audienceAnalyzing better time series with limited human effort is of interest to acade...