Title
|
|
|
|
Hiding global communication latency in the GMRES algorithm on massively parallel machines
| |
Author
|
|
|
|
| |
Abstract
|
|
|
|
In the generalized minimal residual method (GMRES), the global all-to-all communication required in each iteration for orthogonalization and normalization of the Krylov base vectors is becoming a performance bottleneck on massively parallel machines. Long latencies, system noise, and load imbalance cause these global reductions to become very costly global synchronizations. In this work, we propose the use of nonblocking or asynchronous global reductions to hide these global communication latencies by overlapping them with other communications and calculations. A pipelined variation of GMRES is presented in which the result of a global reduction is used only one or more iterations after the communication phase has started. This way, global synchronization is relaxed and scalability is much improved at the expense of some extra computations. The numerical instabilities that inevitably arise due to the typical monomial basis by powering the matrix are reduced and often annihilated by using Newton or Chebyshev bases instead. Our parallel experiments on a medium-sized cluster show significant speedups of the pipelined solvers compared to standard GMRES. An analytical model is used to extrapolate the performance to future exascale systems. |
| |
Language
|
|
|
|
English
| |
Source (journal)
|
|
|
|
SIAM journal on scientific computing. - Philadelphia, Pa
| |
Publication
|
|
|
|
Philadelphia, Pa
:
2013
| |
ISSN
|
|
|
|
1064-8275
| |
DOI
|
|
|
|
10.1137/12086563X
| |
Volume/pages
|
|
|
|
35
:1
(2013)
, p. 48-71
| |
ISI
|
|
|
|
000315575000040
| |
Full text (Publisher's DOI)
|
|
|
|
| |
Full text (open access)
|
|
|
|
| |
|