Programs in Physics & Physical Chemistry
|[Licence| Download | New Version Template] aegy_v1_0.tar.gz(442 Kbytes)|
|Manuscript Title: Parallel Hyperbolic PDE Simulation on Clusters: Cell versus GPU|
|Authors: Scott Rostrup, Hans de Sterck|
|Program title: SWsolver|
|Catalogue identifier: AEGY_v1_0|
Distribution format: tar.gz
|Journal reference: Comput. Phys. Commun. 181(2010)2164|
|Programming language: C, CUDA.|
|Computer: Parallel Computing Clusters. Individual compute nodes may consist of x86 CPU, Cell processor, or x86 CPU with attached NVIDIA GPU accelerator.|
|Operating system: Linux.|
|Has the code been vectorised or parallelized?: Yes. Tested on 1-128 x86 CPU cores, 1-32 Cell Processors, and 1-32 NVIDIA GPUs.|
|RAM: Tested on Problems requiring up to 4 GB per compute node.|
|Keywords: Parallel Computing, Cell Processor, GPU, Hyberbolic PDEs.|
External routines: MPI, CUDA, IBM Cell SDK
Nature of problem:
MPI-parallel simulation of Shallow Water equations using high-resolution 2D hyperbolic equation solver on regular Cartesian grids for x86 CPU, Cell Processor, and NVIDIA GPU using CUDA.
SWsolver provides 3 implementations of a high-resolution 2D Shallow Water equation solver on regular Cartesian grids, for CPU, Cell Processor, and NVIDIA GPU. Each implementation uses MPI to divide work across a parallel computing cluster.
Sub-program numdiff is used for the test run.
The test run provided should run in a few seconds on all architectures. In the results section of the manuscript a comprehensive analysis of performance for different problem sizes and architectures is given.
|Disclaimer | ScienceDirect | CPC Journal | CPC | QUB|