Inverse Preconditioners for Sparse Matrices using CUDA and GPGPU

S. Xu, W. Xue, H.X. Lin

    Research output: Contribution to journalArticleScientificpeer-review

    4 Citations (Scopus)


    The problem of numerical solution of sparse matrix-based linear systems arises from many scientific applications. Iterative solvers and corresponding preconditioning techniques are usually adopted. Due to the the irregularity in memory access patterns and undeterministic branching behavior of sparse matrix-based algorithms, they present unique problems for emerging GPU-based platforms.

    This paper discusses the techniques associated with applying GP-GPU and CUDA to the generation of right-looking approximate inverse preconditioners (AINV) and preconditioned GMRES based on them. Parallel algorithms are designed for generating AINV preconditioners with predefined sparsity pattern or dynamic fill-in control themes. Detailed complexity analysis is provided for dynamic fill-in control in preconditioners. Shared Memory-based optimization yield 200% speedup in the sparse vector inner products and sparse vector updates; while overall speedup compared with parallel CPU-based implementations are 3 times and 6 times, respectively. The preconditioned GMRES iteration based on AINV witnessed 7 to 9 times speedup, due to the high efficiency of matrix-vector product based preconditioning. The developed techniques can also be applied to other Krylov solvers and preconditioners.
    Original languageEnglish
    Pages (from-to)475-500
    Number of pages26
    JournalJournal of Algorithms and Computational Technology
    Issue number3
    Publication statusPublished - 2011


    Dive into the research topics of 'Inverse Preconditioners for Sparse Matrices using CUDA and GPGPU'. Together they form a unique fingerprint.

    Cite this