A series of experiments has been run to examine the floating-point performance of FORTRAN programs on the Flex/32 (Trademark) computer. The experiments are described, and the timing results are presented. The time required to execute a floating-point operation is found to vary considerbaly depending on a number of factors. One factor of particular interest from an algorithm design standpoint is the difference in speed between common memory accesses and local memory accesses. Common memory accesses were found to be slower, and guidelines are given for determinig when it may be cost effective to copy data from common to local memory
Thesis (M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
The use of Force, a parallel, portable FORTRAN on shared memory parallel computers is described. For...
Floating-point computations produce approximate results, which can lead to inaccuracy problems. Exis...
Results are presented of comparative timing tests made by running a typical FORTRAN physics simulati...
A suite of thirteen large Fortran benchmark codes were run on Cray-2 and Cray X-MP supercomputers. T...
A suite of 13 large Fortran benchmark codes were run on a Cray-2 configured with memory pseudo-banki...
The fast Fourier transform (FFT) kernel of the NAS benchmark program has been utilized to experiment...
Applications that get their inputs from sensors are an important and often overlooked application do...
Experiments were conducted at NASA Ames Research Center to define multi-tasking software requirement...
We have developed a performance bounding methodology that explains the performance of loop-dominated...
A guideline for computer programmers who may need to exchange FORTRAN programs between several compu...
It has been shown that FPGAs could outperform high-end microprocessors on floating-point computation...
The methodology is described for converting a large, long-running applications code that executed on...
A methodology for writing parallel programs for shared memory multiprocessors has been formalized as...
This paper addresses the potential speedup achieved by using decimal floating-point hardware, instea...
Thesis (M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
The use of Force, a parallel, portable FORTRAN on shared memory parallel computers is described. For...
Floating-point computations produce approximate results, which can lead to inaccuracy problems. Exis...
Results are presented of comparative timing tests made by running a typical FORTRAN physics simulati...
A suite of thirteen large Fortran benchmark codes were run on Cray-2 and Cray X-MP supercomputers. T...
A suite of 13 large Fortran benchmark codes were run on a Cray-2 configured with memory pseudo-banki...
The fast Fourier transform (FFT) kernel of the NAS benchmark program has been utilized to experiment...
Applications that get their inputs from sensors are an important and often overlooked application do...
Experiments were conducted at NASA Ames Research Center to define multi-tasking software requirement...
We have developed a performance bounding methodology that explains the performance of loop-dominated...
A guideline for computer programmers who may need to exchange FORTRAN programs between several compu...
It has been shown that FPGAs could outperform high-end microprocessors on floating-point computation...
The methodology is described for converting a large, long-running applications code that executed on...
A methodology for writing parallel programs for shared memory multiprocessors has been formalized as...
This paper addresses the potential speedup achieved by using decimal floating-point hardware, instea...
Thesis (M.Eng.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer...
The use of Force, a parallel, portable FORTRAN on shared memory parallel computers is described. For...
Floating-point computations produce approximate results, which can lead to inaccuracy problems. Exis...