We address the parallelization of the LU factorization of hierarchical matrices (-matrices) arising from boundary element methods. Our approach exploits task-parallelism via the OmpSs programming model and runtime, which discovers the data-flow parallelism intrinsic to the operation at execution time, via the analysis of data dependencies based on the memory addresses of the tasks’ operands. This is especially challenging for H-matrices, as the structures containing the data vary in dimension during the execution. We tackle this issue by decoupling the data structure from that used to detect dependencies. Furthermore, we leverage the support for weak operands and early release of dependencies, recently introduced in OmpSs-2, to accelerate t...
Factorization based preconditioning algorithms, most notably incomplete LU (ILU) factorization, have...
In this paper, we investigate how to exploit task-parallelism during the execution of the Cholesky f...
With the emergence of thread-level parallelism as the primary means for continued improvement of per...
We address the parallelization of the LU factorization of hierarchical matrices (-matrices) arising ...
H-matrices offer log-linear storage and computations costs, thanks to a controlled accuracy loss. Th...
International audienceHierarchical matrices (H-matrices) have become important in applications where...
We investigate a parallelization strategy for dense matrix factorization (DMF) algorithms, using Ope...
In this paper we review the technique of hierarchical matrices and put it into the context of black-...
In this paper, we describe and evaluate an extension of the Chameleon library to operate with hierar...
A version of the H-LU factorization is introduced, based on the individual compu-tational tasks occu...
In this work, we consider the reformulation of hierarchical ($\mathcal{H}$) matrix algorithm...
Compression techniques have revolutionized the Boundary Element Method used to solve the Maxwell equ...
We extend a two-level task partitioning previously applied to the inversion of dense matrices via Ga...
In this study, we evaluate two task frameworks with dependencies for important application kernels c...
The number of cores in multicore computers has an irreversible tendency to increase. Also, computers...
Factorization based preconditioning algorithms, most notably incomplete LU (ILU) factorization, have...
In this paper, we investigate how to exploit task-parallelism during the execution of the Cholesky f...
With the emergence of thread-level parallelism as the primary means for continued improvement of per...
We address the parallelization of the LU factorization of hierarchical matrices (-matrices) arising ...
H-matrices offer log-linear storage and computations costs, thanks to a controlled accuracy loss. Th...
International audienceHierarchical matrices (H-matrices) have become important in applications where...
We investigate a parallelization strategy for dense matrix factorization (DMF) algorithms, using Ope...
In this paper we review the technique of hierarchical matrices and put it into the context of black-...
In this paper, we describe and evaluate an extension of the Chameleon library to operate with hierar...
A version of the H-LU factorization is introduced, based on the individual compu-tational tasks occu...
In this work, we consider the reformulation of hierarchical ($\mathcal{H}$) matrix algorithm...
Compression techniques have revolutionized the Boundary Element Method used to solve the Maxwell equ...
We extend a two-level task partitioning previously applied to the inversion of dense matrices via Ga...
In this study, we evaluate two task frameworks with dependencies for important application kernels c...
The number of cores in multicore computers has an irreversible tendency to increase. Also, computers...
Factorization based preconditioning algorithms, most notably incomplete LU (ILU) factorization, have...
In this paper, we investigate how to exploit task-parallelism during the execution of the Cholesky f...
With the emergence of thread-level parallelism as the primary means for continued improvement of per...