A notable characteristic of the scientific computing and machine learning prob-lem domains is the large amount of data to be analyzed and manipulated. Here, carefully crafted parallel algorithms have the potential to make a massive impact on the size of a problem that is deemed tractable. Sparse matrices are a data struc-ture frequently encountered within these domains. Typically, these sparse matri-ces have a non-uniform structure, which can make the task of designing efficient parallel algorithms for sparse matrices much more complex than for their dense counterparts. In this report, we consider multiple ways of parallelizing the most computationally intensive section of a typical sparse matrix algorithm — its inner loop. We compare the t...
Abstract. Sparse matrix-vector multiplication is an important computational kernel that tends to per...
Abstract—Shared-memory systems such as regular desktops now possess enough memory to store large dat...
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel...
Abstract. Large–scale computation on graphs and other discrete struc-tures is becoming increasingly ...
Sparse matrix problems are difficult to parallelize efficiently on distributed memory machines since...
International audienceMany applications in scientific computing process very large sparse matrices o...
International audienceWe discuss efficient shared memory parallelization of sparse matrix computatio...
Matrix factorization is a common task underlying several machine learning applications such as recom...
Vector computers have been extensively used for years in matrix algebra to treat with large dense ma...
Several fine grained parallel algorithms were developed and compared to compute the Cholesky factori...
Many data mining algorithms rely on eigenvalue computations or iterative linear solvers in which the...
In parallel computing environments from multicore systems to cloud computers and supercomputers, dat...
Sparse matrices are first class objects in many VHLLs (very high level languages) used for scientifi...
An efficient algorithm for parallel acquisition of visualization data for large sparse matrices is p...
International audienceIn this paper, we propose a generic method of automatic parallelization for sp...
Abstract. Sparse matrix-vector multiplication is an important computational kernel that tends to per...
Abstract—Shared-memory systems such as regular desktops now possess enough memory to store large dat...
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel...
Abstract. Large–scale computation on graphs and other discrete struc-tures is becoming increasingly ...
Sparse matrix problems are difficult to parallelize efficiently on distributed memory machines since...
International audienceMany applications in scientific computing process very large sparse matrices o...
International audienceWe discuss efficient shared memory parallelization of sparse matrix computatio...
Matrix factorization is a common task underlying several machine learning applications such as recom...
Vector computers have been extensively used for years in matrix algebra to treat with large dense ma...
Several fine grained parallel algorithms were developed and compared to compute the Cholesky factori...
Many data mining algorithms rely on eigenvalue computations or iterative linear solvers in which the...
In parallel computing environments from multicore systems to cloud computers and supercomputers, dat...
Sparse matrices are first class objects in many VHLLs (very high level languages) used for scientifi...
An efficient algorithm for parallel acquisition of visualization data for large sparse matrices is p...
International audienceIn this paper, we propose a generic method of automatic parallelization for sp...
Abstract. Sparse matrix-vector multiplication is an important computational kernel that tends to per...
Abstract—Shared-memory systems such as regular desktops now possess enough memory to store large dat...
Algorithms are often parallelized based on data dependence analysis manually or by means of parallel...