Abstract
The problem of numerical solution of sparse matrix-based linear systems arises from many scientific applications. Iterative solvers and corresponding preconditioning techniques are usually adopted. Due to the the irregularity in memory access patterns and undeterministic branching behavior of sparse matrix-based algorithms, they present unique problems for emerging GPU-based platforms. This paper discusses the techniques associated with applying GP-GPU and CUDA to the generation of right-looking approximate inverse preconditioners (AINV) and preconditioned GMRES based on them. Parallel algorithms are designed for generating AINV preconditioners with predefined sparsity pattern or dynamic fill-in control themes. Detailed complexity analysis is provided for dynamic fill-in control in preconditioners. Shared Memory-based optimization yield 200% speedup in the sparse vector inner products and sparse vector updates; while overall speedup compared with parallel CPU-based implementations are 3 times and 6 times, respectively. The preconditioned GMRES iteration based on AINV witnessed 7 to 9 times speedup, due to the high efficiency of matrix-vector product based preconditioning. The developed techniques can also be applied to other Krylov solvers and preconditioners.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Journal of Algorithms & Computational Technology
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.