z-logo
open-access-imgOpen Access
Accelerating an Adaptive Mesh Refinement Code for Depth‐Averaged Flows Using GPUs
Author(s) -
Qin Xinsheng,
LeVeque Randall J.,
Motley Michael R.
Publication year - 2019
Publication title -
journal of advances in modeling earth systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 3.03
H-Index - 58
ISSN - 1942-2466
DOI - 10.1029/2019ms001635
Subject(s) - computer science , cuda , adaptive mesh refinement , computational science , finite volume method , parallel computing , shallow water equations , correctness , code (set theory) , algorithm , geology , oceanography , physics , set (abstract data type) , mechanics , programming language
Solving the shallow water equations efficiently is critical to the study of natural hazards induced by tsunami and storm surge, since it provides more response time in an early warning system and allows more runs to be done for probabilistic assessment where thousands of runs may be required. Using adaptive mesh refinement speeds up the process by greatly reducing computational demands while accelerating the code using the graphics processing unit (GPU) does so through using faster hardware. Combining both, we present an efficient CUDA implementation of GeoClaw, an open source Godunov‐type high‐resolution finite volume numerical scheme on adaptive grids for shallow water system with varying topography. The use of adaptive mesh refinement and spherical coordinates allows modeling transoceanic tsunami simulation. Numerical experiments on the 2011 Japan tsunami and a local tsunami triggered by a hypothetical M w  7.3 earthquake on the Seattle Fault illustrate the correctness and efficiency of the code, which implements a simplified dimensionally split version of the algorithms. Both numerical simulations are conducted on subregions on a sphere with adaptive grids that adequately resolve the propagating waves. The implementation is shown to be accurate and faster than the original when using Central Processing Units (CPUs) alone. The GPU implementation, when running on a single GPU, is observed to be 3.6 to 6.4 times faster than the original model running in parallel on a 16‐core CPU. Three metrics are proposed to evaluate relative performance of the model, which shows efficient usage of hardware resources.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here