We discuss some techniques for preserving locality of reference in index spaces when mapped to memory units in a distributed memory architecture. In particular, we discuss the use of multidimensional address spaces instead of linearized address spaces, partitioning of irregular grids, and placement of partitions among nodes. We also discuss a set of communication primitives we have found very useful on the Connection Machine systems in implementing scientific and engineering applications. We briefly review some of the techniques used to fully utilize the bandwidth of the binary cube network of the CM-2 and CM-200, and give some performance data from implementations of communication primitives.Engineering and Applied Science
High performance computing (HPC) architectures are specialized machines which can reach their peak p...
This work explores the tradeoffs of the memory system of a new massively parallel multiprocessor in ...
Petascale machines with hundreds of thousands of cores are being built. These machines have varying ...
Massively parallel computing holds the promise of extreme performance. The utility of these systems ...
This article focuses on principles for the design of efficient parallel algorithms for distributed m...
This paper is devoted to the design of communication and memory architectures of massively parallel ...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Abstract: High performance computing (HPC) architectures are specialized machines which can reach th...
This paper describes a technique for improving the data ref-erence locality of parallel programs usi...
We present algorithms for the transportation of data in parallel and distributed systems that would ...
We review a decade\u27s work on message passing MIMD parallel computers in the areas of hardware, so...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Massively parallel computers (MPC) are characterized by the distribution of memory among an ensemble...
Many parallel systems offer a simple view of memory: all storage cells are addressed uniformly. Desp...
We provide performance models for several primitive operations on data structures distributed over m...
High performance computing (HPC) architectures are specialized machines which can reach their peak p...
This work explores the tradeoffs of the memory system of a new massively parallel multiprocessor in ...
Petascale machines with hundreds of thousands of cores are being built. These machines have varying ...
Massively parallel computing holds the promise of extreme performance. The utility of these systems ...
This article focuses on principles for the design of efficient parallel algorithms for distributed m...
This paper is devoted to the design of communication and memory architectures of massively parallel ...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Abstract: High performance computing (HPC) architectures are specialized machines which can reach th...
This paper describes a technique for improving the data ref-erence locality of parallel programs usi...
We present algorithms for the transportation of data in parallel and distributed systems that would ...
We review a decade\u27s work on message passing MIMD parallel computers in the areas of hardware, so...
Massively Parallel Processor systems provide the required computational power to solve most large sc...
Massively parallel computers (MPC) are characterized by the distribution of memory among an ensemble...
Many parallel systems offer a simple view of memory: all storage cells are addressed uniformly. Desp...
We provide performance models for several primitive operations on data structures distributed over m...
High performance computing (HPC) architectures are specialized machines which can reach their peak p...
This work explores the tradeoffs of the memory system of a new massively parallel multiprocessor in ...
Petascale machines with hundreds of thousands of cores are being built. These machines have varying ...