Cargando…
Performance impact of precision reduction in sparse linear systems solvers
It is well established that reduced precision arithmetic can be exploited to accelerate the solution of dense linear systems. Typical examples are mixed precision algorithms that reduce the execution time and the energy consumption of parallel solvers for dense linear systems by factorizing a matrix...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
PeerJ Inc.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8771784/ https://www.ncbi.nlm.nih.gov/pubmed/35111903 http://dx.doi.org/10.7717/peerj-cs.778 |
_version_ | 1784635690176217088 |
---|---|
author | Zounon, Mawussi Higham, Nicholas J. Lucas, Craig Tisseur, Françoise |
author_facet | Zounon, Mawussi Higham, Nicholas J. Lucas, Craig Tisseur, Françoise |
author_sort | Zounon, Mawussi |
collection | PubMed |
description | It is well established that reduced precision arithmetic can be exploited to accelerate the solution of dense linear systems. Typical examples are mixed precision algorithms that reduce the execution time and the energy consumption of parallel solvers for dense linear systems by factorizing a matrix at a precision lower than the working precision. Much less is known about the efficiency of reduced precision in parallel solvers for sparse linear systems, and existing work focuses on single core experiments. We evaluate the benefits of using single precision arithmetic in solving a double precision sparse linear system using multiple cores. We consider both direct methods and iterative methods and we focus on using single precision for the key components of LU factorization and matrix–vector products. Our results show that the anticipated speedup of 2 over a double precision LU factorization is obtained only for the very largest of our test problems. We point out two key factors underlying the poor speedup. First, we find that single precision sparse LU factorization is prone to a severe loss of performance due to the intrusion of subnormal numbers. We identify a mechanism that allows cascading fill-ins to generate subnormal numbers and show that automatically flushing subnormals to zero avoids the performance penalties. The second factor is the lack of parallelism in the analysis and reordering phases of the solvers and the absence of floating-point arithmetic in these phases. For iterative solvers, we find that for the majority of the matrices computing or applying incomplete factorization preconditioners in single precision provides at best modest performance benefits compared with the use of double precision. We also find that using single precision for the matrix–vector product kernels provides an average speedup of 1.5 over double precision kernels. In both cases some form of refinement is needed to raise the single precision results to double precision accuracy, which will reduce performance gains. |
format | Online Article Text |
id | pubmed-8771784 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | PeerJ Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-87717842022-02-01 Performance impact of precision reduction in sparse linear systems solvers Zounon, Mawussi Higham, Nicholas J. Lucas, Craig Tisseur, Françoise PeerJ Comput Sci Algorithms and Analysis of Algorithms It is well established that reduced precision arithmetic can be exploited to accelerate the solution of dense linear systems. Typical examples are mixed precision algorithms that reduce the execution time and the energy consumption of parallel solvers for dense linear systems by factorizing a matrix at a precision lower than the working precision. Much less is known about the efficiency of reduced precision in parallel solvers for sparse linear systems, and existing work focuses on single core experiments. We evaluate the benefits of using single precision arithmetic in solving a double precision sparse linear system using multiple cores. We consider both direct methods and iterative methods and we focus on using single precision for the key components of LU factorization and matrix–vector products. Our results show that the anticipated speedup of 2 over a double precision LU factorization is obtained only for the very largest of our test problems. We point out two key factors underlying the poor speedup. First, we find that single precision sparse LU factorization is prone to a severe loss of performance due to the intrusion of subnormal numbers. We identify a mechanism that allows cascading fill-ins to generate subnormal numbers and show that automatically flushing subnormals to zero avoids the performance penalties. The second factor is the lack of parallelism in the analysis and reordering phases of the solvers and the absence of floating-point arithmetic in these phases. For iterative solvers, we find that for the majority of the matrices computing or applying incomplete factorization preconditioners in single precision provides at best modest performance benefits compared with the use of double precision. We also find that using single precision for the matrix–vector product kernels provides an average speedup of 1.5 over double precision kernels. In both cases some form of refinement is needed to raise the single precision results to double precision accuracy, which will reduce performance gains. PeerJ Inc. 2022-01-17 /pmc/articles/PMC8771784/ /pubmed/35111903 http://dx.doi.org/10.7717/peerj-cs.778 Text en ©2022 Zounon et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, reproduction and adaptation in any medium and for any purpose provided that it is properly attributed. For attribution, the original author(s), title, publication source (PeerJ Computer Science) and either DOI or URL of the article must be cited. |
spellingShingle | Algorithms and Analysis of Algorithms Zounon, Mawussi Higham, Nicholas J. Lucas, Craig Tisseur, Françoise Performance impact of precision reduction in sparse linear systems solvers |
title | Performance impact of precision reduction in sparse linear systems solvers |
title_full | Performance impact of precision reduction in sparse linear systems solvers |
title_fullStr | Performance impact of precision reduction in sparse linear systems solvers |
title_full_unstemmed | Performance impact of precision reduction in sparse linear systems solvers |
title_short | Performance impact of precision reduction in sparse linear systems solvers |
title_sort | performance impact of precision reduction in sparse linear systems solvers |
topic | Algorithms and Analysis of Algorithms |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8771784/ https://www.ncbi.nlm.nih.gov/pubmed/35111903 http://dx.doi.org/10.7717/peerj-cs.778 |
work_keys_str_mv | AT zounonmawussi performanceimpactofprecisionreductioninsparselinearsystemssolvers AT highamnicholasj performanceimpactofprecisionreductioninsparselinearsystemssolvers AT lucascraig performanceimpactofprecisionreductioninsparselinearsystemssolvers AT tisseurfrancoise performanceimpactofprecisionreductioninsparselinearsystemssolvers |