[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1109/CLUSTER.2011.42guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Optimized Non-contiguous MPI Datatype Communication for GPU Clusters: Design, Implementation and Evaluation with MVAPICH2

Published: 26 September 2011 Publication History

Abstract

Data parallel architectures, such as General Purpose Graphics Units (GPGPUs) have seen a tremendous rise in their application for High End Computing. However, data movement in and out of GPGPUs remains the biggest hurdle to overall performance and programmer productivity. Real scientific applications utilize multi-dimensional data. Data in higher dimensions may not be contiguous in memory. In order to improve programmer productivity and to enable communication libraries to optimize non-contiguous data communication, the MPI interface provides MPI data types. Currently, state of the art MPI libraries do not provide native data type support for data that resides in GPU memory. The management of non-contiguous GPU data is a source of productivity and performance loss, because GPU application developers have to manually move the data out of and in to GPUs. In this paper, we present our design for enabling high-performance communication support between GPUs for non-contiguous data types. We describe our innovative approach to improve performance by "offloading" data type packing and unpacking on to a GPU device, and "pipelining" all data transfer stages between two GPUs. Our design is integrated into the popular MVAPICH2 MPI library for InfiniBand, iWARP and RoCE clusters. We perform a detailed evaluation of our design on a GPU cluster with the latest NVIDIA Fermi GPU adapters. The evaluation reveals that the proposed designs can achieve up to 88% latency improvement for vector data type at 4 MB size with micro benchmarks. For Stencil2D application from the SHOC benchmark suite, our design can simplify the data communication in its main loop, reducing the lines of code by 36%. Further, our method can improve the performance of Stencil2D by up to 42% for single precision data set, and 39% for double precision data set. To the best of our knowledge, this is the first such design, implementation and evaluation of non-contiguous MPI data communication for GPU clusters.

Cited By

View all
  • (2023)Evaluating the Viability of LogGP for Modeling MPI Performance with Non-contiguous Datatypes on Modern ArchitecturesProceedings of the 30th European MPI Users' Group Meeting10.1145/3615318.3615326(1-10)Online publication date: 11-Sep-2023
  • (2021)TEMPIProceedings of the 30th International Symposium on High-Performance Parallel and Distributed Computing10.1145/3431379.3460645(95-106)Online publication date: 21-Jun-2021
  • (2019)Network-accelerated non-contiguous memory transfersProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis10.1145/3295500.3356189(1-14)Online publication date: 17-Nov-2019
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Guide Proceedings
CLUSTER '11: Proceedings of the 2011 IEEE International Conference on Cluster Computing
September 2011
613 pages
ISBN:9780769545165

Publisher

IEEE Computer Society

United States

Publication History

Published: 26 September 2011

Author Tags

  1. Cluster
  2. GPGPU
  3. MPI
  4. Non-contiguous

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 16 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2023)Evaluating the Viability of LogGP for Modeling MPI Performance with Non-contiguous Datatypes on Modern ArchitecturesProceedings of the 30th European MPI Users' Group Meeting10.1145/3615318.3615326(1-10)Online publication date: 11-Sep-2023
  • (2021)TEMPIProceedings of the 30th International Symposium on High-Performance Parallel and Distributed Computing10.1145/3431379.3460645(95-106)Online publication date: 21-Jun-2021
  • (2019)Network-accelerated non-contiguous memory transfersProceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis10.1145/3295500.3356189(1-14)Online publication date: 17-Nov-2019
  • (2018)MPI Derived DatatypesProceedings of the 25th European MPI Users' Group Meeting10.1145/3236367.3236378(1-10)Online publication date: 23-Sep-2018
  • (2016)GPU-Aware Non-contiguous Data Movement In Open MPIProceedings of the 25th ACM International Symposium on High-Performance Parallel and Distributed Computing10.1145/2907294.2907317(231-242)Online publication date: 31-May-2016
  • (2016)Designing high performance communication runtime for GPU managed memoryProceedings of the 9th Annual Workshop on General Purpose Processing using Graphics Processing Unit10.1145/2884045.2884050(82-91)Online publication date: 12-Mar-2016
  • (2016)CUDA kernel based collective reduction operations on large-scale GPU clustersProceedings of the 16th IEEE/ACM International Symposium on Cluster, Cloud, and Grid Computing10.1109/CCGrid.2016.111(726-735)Online publication date: 16-May-2016
  • (2015)GPU-Aware Design, Implementation, and Evaluation of Non-blocking Collective BenchmarksProceedings of the 22nd European MPI Users' Group Meeting10.1145/2802658.2802672(1-10)Online publication date: 21-Sep-2015
  • (2015)High performance computing of fiber scattering simulationProceedings of the 8th Workshop on General Purpose Processing using GPUs10.1145/2716282.2716285(90-98)Online publication date: 7-Feb-2015
  • (2014)Energy-efficient collective reduce and allreduce operations on distributed GPUsProceedings of the 14th IEEE/ACM International Symposium on Cluster, Cloud, and Grid Computing10.1109/CCGrid.2014.21(483-492)Online publication date: 26-May-2014
  • Show More Cited By

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media