By Topic

GPU Implementation of Stony Brook University 5-Class Cloud Microphysics Scheme in the WRF

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Jarno Mielikainen ; Space Science and Engineering Center, University of Wisconsin, Madison, WI, USA ; Bormin Huang ; Hung-Lung Allen Huang ; Mitchell D. Goldberg

The Weather Research and Forecasting (WRF) model is a next-generation mesoscale numerical weather prediction system. It is designed to serve the needs of both operational forecasting and atmospheric research for a broad spectrum of applications across scales ranging from meters to thousands of kilometers. Microphysics plays an important role in weather and climate prediction. Microphysics includes explicitly resolved water vapor, cloud, and precipitation processes. Several bulk water microphysics schemes are available within the WRF, with different numbers of simulated hydrometeor classes and methods for estimating their size, fall speeds, distributions and densities. Stony Brook University scheme is a 5-class scheme with riming intensity predicted to account for the mixed-phase processes. In this paper, we develop an efficient Graphics Processing Unit (GPU) based Stony Brook University scheme. The GPU-based Stony Brook University scheme was compared to a CPU-based single-threaded counterpart on a computational domain of 422 × 297 horizontal grid points with 34 vertical levels. The original Fortran code was first rewritten into a standard C code. After that, C code was verified against Fortran code and CUDA C extensions were added for data parallel execution on GPUs. On a single GPU, we achieved a speed-up of 213× with data I/O and 896 × without I/O on NVIDIA GTX 590. Using multiple GPUs, a speed-up of 352 × is achieved with I/O for 4 GPUs. We will also discuss how data I/O will be less cumbersome if we ran the complete WRF model on GPUs.

Published in:

IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing  (Volume:5 ,  Issue: 2 )