[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1109/ICPP.2011.82guideproceedingsArticle/Chapter ViewAbstractPublication PagesConference Proceedingsacm-pubtype
Article

Accelerating Sparse Matrix Vector Multiplication in Iterative Methods Using GPU

Published: 13 September 2011 Publication History

Abstract

Multiplying a sparse matrix with a vector (spmv for short) is a fundamental operation in many linear algebra kernels. Having an efficient spmv kernel on modern architectures such as the GPUs is therefore of principal interest. The computational challenges that spmv poses are significantlydifferent compared to that of the dense linear algebra kernels. Recent work in this direction has focused on designing data structures to represent sparse matrices so as to improve theefficiency of spmv kernels. However, as the nature of sparseness differs across sparse matrices, there is no clear answer as to which data structure to use given a sparse matrix. In this work, we address this problem by devising techniques to understand the nature of the sparse matrix and then choose appropriate data structures accordingly. By using our technique, we are able to improve the performance of the spmv kernel on an Nvidia Tesla GPU (C1060) by a factor of up to80% in some instances, and about 25% on average compared to the best results of Bell and Garland [3] on the standard dataset (cf. Williams et al. SC'07) used in recent literature. We also use our spmv in the conjugate gradient method and show an average 20% improvement compared to using HYB spmv of [3], on the dataset obtained from the The University of Florida Sparse Matrix Collection [9].

Cited By

View all
  • (2020)Efficient Block Algorithms for Parallel Sparse Triangular SolveProceedings of the 49th International Conference on Parallel Processing10.1145/3404397.3404413(1-11)Online publication date: 17-Aug-2020
  • (2017)Sparse Matrix-Vector Multiplication on GPGPUsACM Transactions on Mathematical Software10.1145/301799443:4(1-49)Online publication date: 9-Jan-2017
  • (2017)SpMV and BiCG-Stab optimization for a class of hepta-diagonal-sparse matrices on GPUThe Journal of Supercomputing10.1007/s11227-017-1972-373:9(3761-3795)Online publication date: 1-Sep-2017
  • Show More Cited By
  1. Accelerating Sparse Matrix Vector Multiplication in Iterative Methods Using GPU

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image Guide Proceedings
    ICPP '11: Proceedings of the 2011 International Conference on Parallel Processing
    September 2011
    796 pages
    ISBN:9780769545103

    Publisher

    IEEE Computer Society

    United States

    Publication History

    Published: 13 September 2011

    Author Tags

    1. GPGPU
    2. Iterative Methods
    3. Sparse Matrix Vector Multiplication

    Qualifiers

    • Article

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)0
    • Downloads (Last 6 weeks)0
    Reflects downloads up to 15 Jan 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2020)Efficient Block Algorithms for Parallel Sparse Triangular SolveProceedings of the 49th International Conference on Parallel Processing10.1145/3404397.3404413(1-11)Online publication date: 17-Aug-2020
    • (2017)Sparse Matrix-Vector Multiplication on GPGPUsACM Transactions on Mathematical Software10.1145/301799443:4(1-49)Online publication date: 9-Jan-2017
    • (2017)SpMV and BiCG-Stab optimization for a class of hepta-diagonal-sparse matrices on GPUThe Journal of Supercomputing10.1007/s11227-017-1972-373:9(3761-3795)Online publication date: 1-Sep-2017
    • (2015)A Cholesky preconditioned conjugate gradient algorithm on GPU for the 3D parabolic equationInternational Journal of Computational Science and Engineering10.1504/IJCSE.2015.07349311:4(339-348)Online publication date: 1-Dec-2015
    • (2014)Improving performance by matching imbalanced workloads with heterogeneous platformsProceedings of the 28th ACM international conference on Supercomputing10.1145/2597652.2597675(241-250)Online publication date: 10-Jun-2014
    • (2013)CUDA-enabled Sparse Matrix-Vector Multiplication on GPUs using atomic operationsParallel Computing10.1016/j.parco.2013.09.00539:11(737-750)Online publication date: 1-Nov-2013

    View Options

    View options

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media