There are at least three implications of this work. First, sparse AT Ax should be a basic primitive in sparse matrix libraries, based on its utility to applications and the potential pay-off from automatically tuning it. Second, our upper bound analysis shows that there is an opportunity to apply automatic low-level tuning methods, in the spirit of tuning systems such as ATLAS and PHiPAC for dense linear algebra, to further improve the performance of this kernel. Third, the success of our heuristic provides additional validation of the Sparsity tuning methodology
If A is the (sparse) coefficient matrix of linear equality constraints, for what nonsingular T is fi...
Technology scaling trends have enabled the exponential growth of computing power. However, the perfo...
This report summarizes the progress made as part of a one year lab-directed research and development...
Abstract. This paper presents uniprocessor performance optimizations, automatic tuning techniques, a...
Application performance dominated by a few computational kernels Performance tuning today Vendor-tun...
Sparse kernel performance depends on both the matrix and hardware platform. � Challenges in tuning s...
Abstract. Sparse matrix-vector multiplication is an important computational kernel that tends to per...
This whitepaper describes the programming techniques used to develop an auto-tuning compression sche...
Many data mining algorithms rely on eigenvalue computations or iterative linear solvers in which the...
This whitepaper describes the programming techniques used to develop an auto-tuning compression sche...
SparseM provides some basic R functionality for linear algebra with sparse matrices. Use of the pack...
International audienceWe present a method for automatically selecting optimal implementations of spa...
SparseM provides some basic R functionality for linear algebra with sparse matrices. Use of the pack...
One of the main obstacles to the efficient solution of scientific problems is the problem of tuning ...
Abstract. On many high-speed computers the dense matrix technique is preferable to sparse matrix tec...
If A is the (sparse) coefficient matrix of linear equality constraints, for what nonsingular T is fi...
Technology scaling trends have enabled the exponential growth of computing power. However, the perfo...
This report summarizes the progress made as part of a one year lab-directed research and development...
Abstract. This paper presents uniprocessor performance optimizations, automatic tuning techniques, a...
Application performance dominated by a few computational kernels Performance tuning today Vendor-tun...
Sparse kernel performance depends on both the matrix and hardware platform. � Challenges in tuning s...
Abstract. Sparse matrix-vector multiplication is an important computational kernel that tends to per...
This whitepaper describes the programming techniques used to develop an auto-tuning compression sche...
Many data mining algorithms rely on eigenvalue computations or iterative linear solvers in which the...
This whitepaper describes the programming techniques used to develop an auto-tuning compression sche...
SparseM provides some basic R functionality for linear algebra with sparse matrices. Use of the pack...
International audienceWe present a method for automatically selecting optimal implementations of spa...
SparseM provides some basic R functionality for linear algebra with sparse matrices. Use of the pack...
One of the main obstacles to the efficient solution of scientific problems is the problem of tuning ...
Abstract. On many high-speed computers the dense matrix technique is preferable to sparse matrix tec...
If A is the (sparse) coefficient matrix of linear equality constraints, for what nonsingular T is fi...
Technology scaling trends have enabled the exponential growth of computing power. However, the perfo...
This report summarizes the progress made as part of a one year lab-directed research and development...