Elsevier Science Home
Computer Physics Communications Program Library
Full text online from Science Direct
Programs in Physics & Physical Chemistry
CPC Home

[Licence| Download | New Version Template] aegy_v1_0.tar.gz(442 Kbytes)
Manuscript Title: Parallel Hyperbolic PDE Simulation on Clusters: Cell versus GPU
Authors: Scott Rostrup, Hans de Sterck
Program title: SWsolver
Catalogue identifier: AEGY_v1_0
Distribution format: tar.gz
Journal reference: Comput. Phys. Commun. 181(2010)2164
Programming language: C, CUDA.
Computer: Parallel Computing Clusters. Individual compute nodes may consist of x86 CPU, Cell processor, or x86 CPU with attached NVIDIA GPU accelerator.
Operating system: Linux.
Has the code been vectorised or parallelized?: Yes. Tested on 1-128 x86 CPU cores, 1-32 Cell Processors, and 1-32 NVIDIA GPUs.
RAM: Tested on Problems requiring up to 4 GB per compute node.
Keywords: Parallel Computing, Cell Processor, GPU, Hyberbolic PDEs.
Classification: 12.

External routines: MPI, CUDA, IBM Cell SDK

Nature of problem:
MPI-parallel simulation of Shallow Water equations using high-resolution 2D hyperbolic equation solver on regular Cartesian grids for x86 CPU, Cell Processor, and NVIDIA GPU using CUDA.

Solution method:
SWsolver provides 3 implementations of a high-resolution 2D Shallow Water equation solver on regular Cartesian grids, for CPU, Cell Processor, and NVIDIA GPU. Each implementation uses MPI to divide work across a parallel computing cluster.

Additional comments:
Sub-program numdiff is used for the test run.

Running time:
The test run provided should run in a few seconds on all architectures. In the results section of the manuscript a comprehensive analysis of performance for different problem sizes and architectures is given.