The recent dramatic progress in machine learning is partially attributed to the availability of high-performant computers and development tools. The accelerated linear algebra (XLA) compiler is one such tool that automatically optimises array operations (mostly fusion to reduce memory operations) and compiles the optimised operations into high-performant programs specific to target computing platforms. Like machine-learning models, numerical models are often expressed in array operations, and thus their performance can be boosted by XLA. This study is the first of its kind to examine the efficiency of XLA for numerical models, and the efficiency is examined stringently by comparing its performance with that of optimal implementations. Two s...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
Computers are powerful tools which perform fast, accurate calculations over huge sets of data. Howev...
. In this paper we explore the characteristics of numerically intensive programs and explore their e...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
This thesis describes novel techniques and test implementations for optimizing numerically intensive...
Machine Learning involves analysing large sets of training data to make predictions and decisions to...
Machine learning algorithms are complex to model on hardware. This is due to the fact that these alg...
International audienceIn this work, numerical algebraic operations are performed by using several li...
As Central Processing Units (CPUs) and Graphical Processing Units (GPUs) get progressively better, d...
This dissertation incorporates two research projects: performance modeling and prediction for dense ...
Dense linear algebra(DLA) is one of the most seven important kernels in high performance computing. ...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
Computers are powerful tools which perform fast, accurate calculations over huge sets of data. Howev...
. In this paper we explore the characteristics of numerically intensive programs and explore their e...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
The recent dramatic progress in machine learning is partially attributed to the availability of high...
This thesis describes novel techniques and test implementations for optimizing numerically intensive...
Machine Learning involves analysing large sets of training data to make predictions and decisions to...
Machine learning algorithms are complex to model on hardware. This is due to the fact that these alg...
International audienceIn this work, numerical algebraic operations are performed by using several li...
As Central Processing Units (CPUs) and Graphical Processing Units (GPUs) get progressively better, d...
This dissertation incorporates two research projects: performance modeling and prediction for dense ...
Dense linear algebra(DLA) is one of the most seven important kernels in high performance computing. ...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
Computers are powerful tools which perform fast, accurate calculations over huge sets of data. Howev...
. In this paper we explore the characteristics of numerically intensive programs and explore their e...