thesisScientific libraries are written in a general way in anticipation of a variety of use cases that reduce optimization opportunities. Significant performance gains can be achieved by specializing library code to its execution context: the application in which it is invoked, the input data set used, the architectural platform and its backend compiler. Such specialization is not typically done because it is time-consuming, leads to nonportable code and requires performance-tuning expertise that application scientists may not have. Tool support for library specialization in the above context could potentially reduce the extensive under-standing required while significantly improving performance, code reuse and portability. In this work, we...
This dissertation studies the sources of poor performance in scientific computing codes based on par...
Even parts of a program that are sequential or just inherently difficult to parallelize can be optim...
International audienceThe quality of compiler-optimized code for high-performance applications lags ...
Recent years have witnessed a tremendous surge of interest in accelerating sparse linear algebra app...
dissertationSparse matrix codes are found in numerous applications ranging from iterative numerical ...
International audienceThe ever growing complexity and scale of parallel architectures imposes to rew...
The rise of cloud computing and deep machine learning in recent years have led to a tremendous growt...
Changes and Enhancements for Release 2.0: 4 papers have been added to SparseLab 200: "Fast Solution ...
Scientific applications are computationally intensive and require expensive HPC resources. Optimizin...
The objective of this research is to improve the performance of sparse problems that have a wide ran...
thesisThe advent of the era of cheap and pervasive many-core and multicore parallel sys-tems has hig...
In this thesis, I explore an approach called "active libraries". These are libraries that take part...
As we approach the Exascale computing era, disruptive changes in the software landscape are required...
Portability, an oftentimes sought-after goal in scientific applications, confers a number of possibl...
Progress in Machine Learning is being driven by continued growth in model size, training data and al...
This dissertation studies the sources of poor performance in scientific computing codes based on par...
Even parts of a program that are sequential or just inherently difficult to parallelize can be optim...
International audienceThe quality of compiler-optimized code for high-performance applications lags ...
Recent years have witnessed a tremendous surge of interest in accelerating sparse linear algebra app...
dissertationSparse matrix codes are found in numerous applications ranging from iterative numerical ...
International audienceThe ever growing complexity and scale of parallel architectures imposes to rew...
The rise of cloud computing and deep machine learning in recent years have led to a tremendous growt...
Changes and Enhancements for Release 2.0: 4 papers have been added to SparseLab 200: "Fast Solution ...
Scientific applications are computationally intensive and require expensive HPC resources. Optimizin...
The objective of this research is to improve the performance of sparse problems that have a wide ran...
thesisThe advent of the era of cheap and pervasive many-core and multicore parallel sys-tems has hig...
In this thesis, I explore an approach called "active libraries". These are libraries that take part...
As we approach the Exascale computing era, disruptive changes in the software landscape are required...
Portability, an oftentimes sought-after goal in scientific applications, confers a number of possibl...
Progress in Machine Learning is being driven by continued growth in model size, training data and al...
This dissertation studies the sources of poor performance in scientific computing codes based on par...
Even parts of a program that are sequential or just inherently difficult to parallelize can be optim...
International audienceThe quality of compiler-optimized code for high-performance applications lags ...