International audienceWe present algorithms for real and complex dot product and matrix multiplication in arbitrary-precision floating-point and ball arithmetic. A low-overhead dot product is implemented on the level of GMP limb arrays; it is about twice as fast as previous code in MPFR and Arb at precision up to several hundred bits. Up to 128 bits, it is 3-4 times as fast, costing 20-30 cycles per term for floating-point evaluation and 40-50 cycles per term for balls. We handle large matrix multiplications even more efficiently via blocks of scaled integer matrices. The new methods are implemented in Arb and significantly speed up polynomial operations and linear algebra
International audienceDot products (also called sums of products) are ubiquitous in matrix computati...
Due to non-associativity of floating-point operations and dynamic scheduling on parallel architectur...
AbstractThe complexity of matrix multiplication has attracted a lot of attention in the last forty y...
National audienceOn modern multi-core, many-core, and heterogeneous architectures, floating-point co...
International audienceOn modern multi-core, many-core, and heterogeneous architectures, floating-poi...
International audienceThe straightforward implementation of interval matrix product suf- fers from p...
The complexity of matrix multiplication (hereafter MM) has been intensively studied since 1969, when...
International audienceSome important computational problems must use a floating-point (FP) precision...
International audienceThis paper presents some work in progress on the development of fast and accur...
International audienceDue to non-associativity of floating-point operations and dynamic schedu...
Can post-Schönhage–Strassen multiplication algorithms be competitive in practice for large input siz...
Algorithms for summation and dot product of floating point numbers are presented which are fast in t...
AbstractSeveral different techniques and softwares intend to improve the accuracy of results compute...
Abstract—This paper is concerned with an accurate computation of matrix multiplication, where compon...
International audienceWe present a fast algorithm together with its low-level implementation of corr...
International audienceDot products (also called sums of products) are ubiquitous in matrix computati...
Due to non-associativity of floating-point operations and dynamic scheduling on parallel architectur...
AbstractThe complexity of matrix multiplication has attracted a lot of attention in the last forty y...
National audienceOn modern multi-core, many-core, and heterogeneous architectures, floating-point co...
International audienceOn modern multi-core, many-core, and heterogeneous architectures, floating-poi...
International audienceThe straightforward implementation of interval matrix product suf- fers from p...
The complexity of matrix multiplication (hereafter MM) has been intensively studied since 1969, when...
International audienceSome important computational problems must use a floating-point (FP) precision...
International audienceThis paper presents some work in progress on the development of fast and accur...
International audienceDue to non-associativity of floating-point operations and dynamic schedu...
Can post-Schönhage–Strassen multiplication algorithms be competitive in practice for large input siz...
Algorithms for summation and dot product of floating point numbers are presented which are fast in t...
AbstractSeveral different techniques and softwares intend to improve the accuracy of results compute...
Abstract—This paper is concerned with an accurate computation of matrix multiplication, where compon...
International audienceWe present a fast algorithm together with its low-level implementation of corr...
International audienceDot products (also called sums of products) are ubiquitous in matrix computati...
Due to non-associativity of floating-point operations and dynamic scheduling on parallel architectur...
AbstractThe complexity of matrix multiplication has attracted a lot of attention in the last forty y...