[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Logo PTI Logo FedCSIS

Proceedings of the 18th Conference on Computer Science and Intelligence Systems

Annals of Computer Science and Information Systems, Volume 35

Runge-Kutta Method and WSM6 Microphysics for Weather Prediction on Hybrid Parallel Platform

, ,

DOI: http://dx.doi.org/10.15439/2023F7199

Citation: Proceedings of the 18th Conference on Computer Science and Intelligence Systems, M. Ganzha, L. Maciaszek, M. Paprzycki, D. Ślęzak (eds). ACSIS, Vol. 35, pages 11371141 ()

Full text

Abstract. In Numerical Weather Prediction models we need to model the dynamics of the atmosphere (pressure, temperature, wind, water vapor) and the physical variables (clouds, radiation, precipitation, and others) that take place in a moment. To solve the complex set of model equations on computers different numerical methods can be employed. In grid point models the temporal evolution of the model variables are calculated in a three-dimensional spatial grid which covers the atmosphere from the surface up to a given model top. Most important numerical weather prediction models include two import routine namely Runge-Kutta numerical method and microphysics scheme WSM6. They have been used to obtain relevant information about rain, snow, temperature and others. The Runge-Kutta numerical method is an important routine and it is a very time-consuming task of the model related to the dynamical core, that is responsible for integration in time. Another important routine is the WSM6 microphysics. It is responsible to calculates several hydrometeors variables and is the most consuming time routine in the entire process. Due to them importance, several parallel methods have been proposed to the problem in order to gain time performance. This paper describes the improvement of the computational performance of the Runge-Kutta method and WSM6 micro- physics by exploiting multi-level parallelism using CUDA-based Graphics Processing Unit (GPU) on hybrid parallel platform. We applied pipeline parallelism technique, workload balancing and asynchronous data exchange strategy each demonstrating be useful to improve performance. Our experimental tests show that we achieved speedup ranging from 10 to 39.3 for the Runge-Kutta method and from 5.68 to 26.39 for the WSM6 microphysics when compared to a 12- thread CPU. The pipeline parallelism and asynchronous data exchange strategies improved performance by up to 37\%. We also performed a study about the accuracy of the proposed implementation with good results.

References

  1. M. R. Petersen, X. S. Asay-Davis, D. W. Jacobsen, M. E. Maltrud, T. D. Ringler, L. Van Roekel, C. Veneziani, and P. J. Wolfram, Jr., “MPAS-Ocean model user’s guide version 6.0,” OSTI, Tech. Rep., 4 2018.
  2. S. R. Freitas, J. Panetta, K. M. Longo, L. F. Rodrigues, D. S. Moreira et al., “The brazilian developments on the regional atmospheric modeling system (BRAMS 5.2): an integrated environmental model tuned for tropical areas,” Geosci. Model Dev., vol. 10, pp. 189–222, 2017.
  3. W. Skamarock, J. Klemp, M. Duda, L. Fowler, S. Park, and T. Ringler, “A multiscale nonhydrostatic atmospheric model using centroidal voronoi tesselations and c-grid staggering,” Monthly Weather Review, vol. 140, no. 9, pp. 3090–3105, Sep. 2012.
  4. Z. Ma, Q. Liu, C. Zhao, X. Shen, Y. Wang, J. H. Jiang, Z. Li, and Y. Yung, “Application and evaluation of an explicit prognostic cloud-cover scheme in grapes global forecast system,” Journal of Advances in Modeling Earth Systems, vol. 10, no. 3, pp. 652–667, 2018.
  5. W. C. Skamarock, J. B. Klemp, J. Dudhia, D. O. Gill, M. Barker, K. G. Duda, X. Y. Huang, W. Wang, and J. G. Powers, “A description of the Advanced Research WRF Version 3,” National Center for Atmospheric Research, Tech. Rep., 2008.
  6. M. Korch and T. Rauber, “Comparison of parallel implementations of runge-kutta solvers: Message passing vs. threads,” in Parallel Computing - Advances in Parallel Computing, G. Joubert, W. Nagel, F. Peters, and W. Walter, Eds. North-Holland, 2004, vol. 13, pp. 209–216.
  7. L. Murray, “GPU acceleration of runge-kutta integrators,” IEEE Transactions on Parallel and Distributed Systems, vol. 23, pp. 94–101, 2012.
  8. C. Chantrapornchai, S. Dainpraprai, and O. Wongtaweesap, “On the Computer Simulation of Microparticles Capture in Magnetic Filters using OpenMP,” International Journal of Computer Applications, vol. 51, no. 14, pp. 23–30, Aug. 2012.
  9. M. S. Wo, R. Gobithaasan, and K. Miura, “GPU acceleration of runge kutta-fehlberg and its comparison with dormand-prince method,” AIP Conference Proceedings, vol. 1605, pp. 16–21, 2014.
  10. B. Bylina and J. Potiopa, “Explicit fourth-order runge—-kutta method on intel xeon phi coprocessor,” International Journal of Parallel Programming, vol. 45, no. 5, p. 1073–1090.
  11. M. Huang, B. Huang, L. Gu, H. Huang, and M. Goldberg, “Parallel GPU architecture framework for the WRF single moment 6-class microphysics scheme,” Computers & Geosciences, vol. 83, 06 2015.
  12. J. Y. Kim, J.-S. Kang, and M. Joh, “GPU acceleration of MPAS microphysics WSM6 using OpenACC directives: Performance and verification,” Computers and Geosciences, vol. 146, p. 104627, Jan. 2021.
  13. H. C. da Silva, M. A. Stefanes, and V. Capistrano, “OpenACC Multi-GPU approach for WSM6 microphysics,” in 2021 IEEE 28th Interna- tional Conference on High Performance Computing, Data, and Analytics (HiPC), 2021, pp. 382–387.
  14. J. B. Klemp, W. C. Skamarock, and J. Dudhia, “Conservative Split-Explicit Time Integration Methods for the Compressible Nonhydrostatic Equations,” Monthly Weather Review, vol. 135, pp. 2897–2913, 2007.
  15. S. Hong and J. Lim, “The WRF Single-Moment 6-Class Microphysics Scheme (WSM6),” J. Korean Meteor. Soc., vol. 42, pp. 129–151, 2006.
  16. Message Passing Interface Forum, MPI: A Message-Passing Interface Standard Version 4.0, Jun. 2021.
  17. R. Chandra, L. Dagum, D. Kohr, R. Menon, D. Maydan, and J. McDonald, Parallel programming in OpenMP. Morgan kaufmann, 2001.
  18. S. Cook, CUDA Programming: A Developer’s Guide to Parallel Computing with GPUs. Elsevier Science, 2012.
  19. W. Collins, P. Rasch, B. Boville, J. McCaa, D. Williamson, J. Kiehl, B. Briegleb, C. Bitz, S. Lin, M. Zhang, and Y. Dai, “Description of the NCAR Community Atmosphere Model (CAM 3.0),” Tech. Rep., 2004.
  20. S.-Y. Hong, Y. Noh, and J. Dudhia, “A New Vertical Diffusion Package with an Explicit Treatment of Entrainment Processes,” Monthly Weather Review, vol. 134, no. 9, pp. 2318–2341, Sep. 2006.
  21. J. S. Kain, “The Kain–Fritsch Convective Parameterization: An Update,” Journal of Applied Meteorology, vol. 43, no. 1, pp. 170 – 181, 2004.