z-logo
Premium
A massively parallel GPU‐accelerated model for analysis of fully nonlinear free surface waves
Author(s) -
EngsigKarup A. P.,
Madsen Morten G.,
Glimberg Stefan L.
Publication year - 2011
Publication title -
international journal for numerical methods in fluids
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.938
H-Index - 112
eISSN - 1097-0363
pISSN - 0271-2091
DOI - 10.1002/fld.2675
Subject(s) - computer science , massively parallel , parallel computing , computational science , graphics processing unit , single core , scalability , thread (computing) , cuda , fortran , nonlinear system , grid , multigrid method , multi core processor , physics , geometry , mathematics , quantum mechanics , database , operating system
SUMMARY We implement and evaluate a massively parallel and scalable algorithm based on a multigrid preconditioned Defect Correction method for the simulation of fully nonlinear free surface flows. The simulations are based on a potential model that describes wave propagation over uneven bottoms in three space dimensions and is useful for fast analysis and prediction purposes in coastal and offshore engineering. A dedicated numerical model based on the proposed algorithm is executed in parallel by utilizing affordable modern special purpose graphics processing unit (GPU). The model is based on a low‐storage flexible‐order accurate finite difference method that is known to be efficient and scalable on a CPU core (single thread). To achieve parallel performance of the relatively complex numerical model, we investigate a new trend in high‐performance computing where many‐core GPUs are utilized as high‐throughput co‐processors to the CPU. We describe and demonstrate how this approach makes it possible to do fast desktop computations for large nonlinear wave problems in numerical wave tanks (NWTs) with close to 50/100 million total grid points in double/single precision with 4 GB global device memory available. A new code base has been developed in C++ and compute unified device architecture C and is found to improve the runtime more than an order in magnitude in double precision arithmetic for the same accuracy over an existing CPU (single thread) Fortran 90 code when executed on a single modern GPU. These significant improvements are achieved by carefully implementing the algorithm to minimize data‐transfer and take advantage of the massive multi‐threading capability of the GPU device. Copyright © 2011 John Wiley & Sons, Ltd.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here