International audienceOn modern architectures, the performance of 32-bit operations is often at least twice as fast as the performance of 64-bit operations. By using a combination of 32-bit and 64-bit floating point arithmetic, the performance of many dense and sparse linear algebra algorithms can be significantly enhanced while maintaining the 64-bit accuracy of the resulting solution. The approach presented here can apply not only to conventional processors but also to other technologies such as Field Programmable Gate Arrays (FPGA), Graphical Processing Units (GPU), and the STI Cell BE processor. Results on modern processor architectures and the STI Cell BE are presented
The slowing pace of commodity microprocessor performance improvements combined with ever-increasing ...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
We present two designs (I and II) for IEEE 754 double precision floating point matrix multiplication...
International audienceOn modern architectures, the performance of 32-bit operations is often at leas...
International audienceBy using a combination of 32-bit and 64-bit floating point arithmetic, the per...
By using a combination of 32-bit and 64-bit floating point arithmetic, the per-formance of many dens...
International audienceBy using a combination of 32-bit and 64-bit floating point arithmetic, the per...
Recent versions of microprocessors exhibit performance characteristics for 32 bit floating point ari...
Sparse linear algebra arises in a wide variety of computational disciplines, including medical imagi...
Floating-point computing with more than one TFLOP of peak performance is already a reality in recent...
Part 4: Architecture and HardwareInternational audienceMatrix computing plays a vital role in many s...
We present several algorithms to compute the solution of a linear system of equations on a graphics ...
UnrestrictedThe large capacity of field programmable gate arrays (FPGAs) has prompted researchers to...
The STI CELL processor introduces pioneering solutions in processor architecture. At the same time i...
Basic Linear Algebra Subprograms (BLAS) and Linear Algebra Package (LAPACK) form basic building bloc...
The slowing pace of commodity microprocessor performance improvements combined with ever-increasing ...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
We present two designs (I and II) for IEEE 754 double precision floating point matrix multiplication...
International audienceOn modern architectures, the performance of 32-bit operations is often at leas...
International audienceBy using a combination of 32-bit and 64-bit floating point arithmetic, the per...
By using a combination of 32-bit and 64-bit floating point arithmetic, the per-formance of many dens...
International audienceBy using a combination of 32-bit and 64-bit floating point arithmetic, the per...
Recent versions of microprocessors exhibit performance characteristics for 32 bit floating point ari...
Sparse linear algebra arises in a wide variety of computational disciplines, including medical imagi...
Floating-point computing with more than one TFLOP of peak performance is already a reality in recent...
Part 4: Architecture and HardwareInternational audienceMatrix computing plays a vital role in many s...
We present several algorithms to compute the solution of a linear system of equations on a graphics ...
UnrestrictedThe large capacity of field programmable gate arrays (FPGAs) has prompted researchers to...
The STI CELL processor introduces pioneering solutions in processor architecture. At the same time i...
Basic Linear Algebra Subprograms (BLAS) and Linear Algebra Package (LAPACK) form basic building bloc...
The slowing pace of commodity microprocessor performance improvements combined with ever-increasing ...
Field Programmable Gate Arrays (FPGAs) enable powerful performance acceleration for scientific compu...
We present two designs (I and II) for IEEE 754 double precision floating point matrix multiplication...