scispace - formally typeset
Search or ask a question
Journal ArticleDOI

The Tortoise and the Hare restart GMRES

Mark Embree1
01 Jan 2003-Siam Review (Society for Industrial and Applied Mathematics)-Vol. 45, Iss: 2, pp 259-266
TL;DR: Two simple examples are presented where GM RES(1) converges exactly in three iterations, while GMRES(2) stagnates, revealing that GMRES (1) convergence can be extremely sensitive to small changes in the initial residual.
Abstract: When solving large nonsymmetric systems of linear equations with the restarted GMRES algorithm, one is inclined to select a relatively large restart parameter in the hope of mimicking the full GMRES process. Surprisingly, cases exist where small values of the restart parameter yield convergence in fewer iterations than larger values. Here, two simple examples are presented where GMRES(1) converges exactly in three iterations, while GMRES(2) stagnates. One of these examples reveals that GMRES(1) convergence can be extremely sensitive to small changes in the initial residual.

Content maybe subject to copyright    Report

THE TORTOISE AND THE HARE RESTART GMRES
MARK EMBREE
Abstrat.
When solving large nonsymmetri systems of linear equations with the restarted
GMRES algorithm, one is inlined to selet a relatively large restart parameter in the hop e of
mimiking the full GMRES pro ess. Surprisingly, ases exist where small values of the restart
parameter yield onvergene in fewer iterations than larger values. Here, two simple examples are
presented where GMRES(1) onverges exatly in three iterations, while GMRES(2) stagnates. One
of these examples reveals that GMRES(1) onvergene an b e extremely sensitive to small hanges
in the initial residual.
Key words.
Restarted GMRES, Krylov subspae methods.
AMS sub jet lassiations.
65F10, 37N30
1. Intro dution.
GMRES is an iterative method for solving large nonsymmetri
systems of linear equations,
Ax
=
b
[8℄. Throughout siene and engineering, this
algorithm and its variants routinely solve problems with millions of degrees of freedom.
Its p opularity is rooted in an optimality ondition: At the
k
th iteration, GMRES
omputes the solution estimate
x
k
that minimizes the Eulidean norm of the residual
r
k
=
Ax
k
b
over a subspae of dimension
k
,
k
r
k
k
= min
p
2
P
k
p
(0)=1
k
p
(
A
)
r
0
k
;
(1.1)
where
P
k
denotes those polynomials with degree not exeeding
k
, and
r
0
=
b
Ax
0
is the initial residual. As eah iteration enlarges the minimizing subspae, the residual
norm dereases monotonially.
GMRES optimality omes at a ost, however, sine eah iteration demands both
more arithmeti and memory than the one b efore it. A standard work-around is
to restart the pro ess after some xed number of iterations,
m
. The resulting algo-
rithm, GMRES(
m
), uses the approximate solution
x
m
as the initial guess for a new
run of GMRES, ontinuing this proess until onvergene. The global optimality of
the original algorithm is lost, so although the residual norms remain monotoni, the
restarted pro ess an stagnate with a non-zero residual, failing to ever onverge [8℄.
Sine GMRES(
m
) enfores loal optimality on
m
-dimensional spaes, one antiipates
that inreasing
m
will yield onvergene in fewer iterations. Many pratial examples
onrm this intuition.
We denote the
k
th residual of GMRES(
m
) by
r
(
m
)
k
. To b e preise, one yle
between restarts of GMRES(
m
) is ounted as
m
individual iterations. Conventionally,
then, one exp ets
k
r
(
m
)
k
k k
r
(
`
)
k
k
for
` < m
. Indeed, this must be true when
k
m
.
Surprisingly, inreasing the restart parameter sometimes leads to
slower
onver-
gene:
k
r
(
m
)
k
k
>
k
r
(
`
)
k
k
for
` < m < k
. The author enountered this phenomenon
while solving a disretized onvetion-diusion equation desribed in [4℄. In unpub-
lished exp eriments, de Sturler [1 and Walker and Watson [11 observed similar b e-
havior arising in pratial appliations. One wonders, how muh smaller than
k
r
(
m
)
k
k
might
k
r
(
`
)
k
k
be? The smallest possible ases ompare GMRES(1) to GMRES(2) for
3-by-3 matries. Eiermann, Ernst, and Shneider present suh an example for whih
Oxford University Computing Laboratory, Wolfson Building, Parks Road, Oxford OX1 3QD,
United Kingdom (mark.embreeomlab.ox.a.uk). Supp orted by UK Engineering and Physial Si-
enes Researh Counil Grant GR/M12414.
1

2
MARK EMBREE
k
r
(1)
4
k
=
k
r
(2)
4
k
= 0
:
2154
: : :
[2, pp. 284{285℄. Otherwise, the phenomenon we desrib e
has apparently reeived little attention in the literature.
The purp ose of this artile is twofold. First, we desribe a pair of extreme ex-
amples where GMRES(1) onverges exatly at the third iteration, while GMRES(2)
seems to never onverge. The seond example leads to our seond p oint: Small p er-
turbations to the initial residual an dramatially alter the onvergene b ehavior of
GMRES(1).
2. First Example.
Consider using restarted GMRES to solve
Ax
=
b
for
A
=
0
1 1 1
0 1 3
0 0 1
1
A
;
b
=
0
2
4
1
1
A
:
(2.1)
Taking
x
0
=
0
yields the initial residual
r
0
=
b
. Using the fat that
A
and
r
0
are
real, we an derive expliit formulas for GMRES(1) and GMRES(2) diretly from the
GMRES optimality ondition (1.1). The reurrene for GMRES(1),
r
(1)
k
+1
=
r
(1)
k
r
(1)T
k
Ar
(1)
k
r
(1)T
k
A
T
Ar
(1)
k
Ar
(1)
k
;
(2.2)
was studied as early as the 1950s [3,
x
71℄,[7℄. For the
A
and
r
0
=
b
dened in (2.1),
this iteration onverges
exatly
at the third step:
r
(1)
1
=
0
3
3
0
1
A
;
r
(1)
2
=
0
3
0
0
1
A
;
r
(1)
3
=
0
0
0
0
1
A
:
Expressions for one GMRES(2) yle an likewise b e derived using elementary alu-
lus. The up dated residual takes the form
r
(2)
k
+2
=
p
(
A
)
r
(2)
k
, where
p
(
z
) = 1 +
z
+
z
2
is a quadrati whose o eÆients
=
(
A
;
r
(2)
k
) and
=
(
A
;
r
(2)
k
) are given by
=
(
r
(2)T
k
AAr
(2)
k
)(
r
(2)T
k
A
T
AAr
(2)
k
)
(
r
(2)T
k
Ar
(2)
k
)(
r
(2)T
k
A
T
A
T
AAr
(2)
k
)
(
r
(2)T
k
A
T
Ar
(2)
k
)(
r
(2)T
k
A
T
A
T
AAr
(2)
k
)
(
r
(2)T
k
A
T
AAr
(2)
k
)(
r
(2)T
k
A
T
AAr
(2)
k
)
;
=
(
r
(2)T
k
Ar
(2)
k
)(
r
(2)T
k
A
T
AAr
(2)
k
)
(
r
(2)T
k
AAr
(2)
k
)(
r
(2)T
k
A
T
Ar
(2)
k
)
(
r
(2)T
k
A
T
Ar
(2)
k
)(
r
(2)T
k
A
T
A
T
AAr
(2)
k
)
(
r
(2)T
k
A
T
AAr
(2)
k
)(
r
(2)T
k
A
T
AAr
(2)
k
)
:
Exeuting GMRES(2) on the matrix and right hand side (2.1) reveals
r
(2)
1
=
0
3
3
0
1
A
;
r
(2)
2
=
1
2
0
3
0
3
1
A
;
r
(2)
3
=
1
28
0
24
27
33
1
A
;
r
(2)
4
=
1
122
0
81
108
162
1
A
:
The inferiority of GMRES(2) ontinues well b eyond the fourth iteration. For example:
k
k
r
(2)
k
k
=
k
r
0
k
5 0.376888290025532. . .
10 0.376502488858910. . .
15 0.376496927936533. . .
20 0.376496055944867. . .
25 0.376495995285626. . .
30 0.376495984909087. . .

RESTARTED GMRES
3
k
r
(
m
)
k
k
k
r
0
k
iteration,
k
GMRES(1)
GMRES(2)
0 5 10 15 20 25 30
10
0
10
5
10
10
10
15
Fig. 1
.
Convergene urves for GMRES
(1)
and GMRES
(2)
applied to
(2.1)
with
x
0
=
0
.
The entire onvergene urve for the rst thirty iterations is shown in Figure 1, based
on p erforming GMRES(2) in exat arithmeti using Mathematia.
The partiular value of
b
(and thus
r
0
) studied ab ove is exeptional, as it is
unusual for GMRES(1) to onverge exatly in three iterations. Remarkably, though,
GMRES(1) maintains sup eriority over GMRES(2) for a wide range of initial residuals.
For this matrix
A
, GMRES(2) onverges exatly in one yle for any initial residual
with zero in the third omp onent, so we restrit attention to residuals normalized to
the form
r
0
= (
; ;
1)
T
. Figure 2 indiates that GMRES(2) makes little progress for
most suh residuals, while GMRES(1) onverges to high auray for the vast ma jor-
ity of these
r
0
values. The olor in eah plot reets the magnitude of
k
r
(
m
)
100
k
=
k
r
0
k
:
Blue indiates satisfatory onvergene, while red signals little progress in one hun-
dred iterations. (To ensure this data's delity, we performed these omputations in
both double and quadruple preision arithmeti; dierenes b etween the two were
negligible.)
To gain an appreiation for the dynamis behind Figure 2, we rst examine the
ation of a single GMRES(1) step. From (2.2) it is lear that GMRES(1) will om-
pletely stagnate only when
r
T
0
Ar
0
= 0. For the matrix
A
speied in (2.1) and
r
0
= (
; ;
1)
T
, this ondition redues to
2
+
+
2
+
+ 3
+ 1 = 0
;
(2.3)
the equation for an oblique ellipse in the (
;
) plane.
Now writing
r
(1)
k
= (
; ;
1)
T
, onsider the map
r
(1)
k
7!
s
(1)
k
+1
that pro jets
r
(1)
k
+1
into the (
;
) plane,
s
(1)
k
+1
= (
r
(1)
k
+1
)
1
3
0
(
r
(1)
k
+1
)
1
(
r
(1)
k
+1
)
2
1
A
;

4
MARK EMBREE
10
5 0 5 10
10
5
0
5
10
10
5 0 5 10
10
5
0
5
10
10
15
10
10
10
5
10
0
Fig. 2
.
Convergene of GMRES
(1) (
left
)
and GMRES
(2) (
right
)
for the matrix in
(2.1)
over
a range of initial residuals of the form
r
0
= (
; ;
1)
T
. The olor indiates
k
r
(
m
)
100
k
=
k
r
0
k
on a loga-
rithmi sale: blue regions orrespond to initial residuals that onverge satisfatorily, while the red
regions show residuals that stagnate or onverge very slow ly.
where (
r
(1)
k
+1
)
j
denotes the
j
th entry of
r
(1)
k
+1
, whih itself is derived from
r
(1)
k
via (2.2).
For the present example, we have
s
(1)
k
+1
=
0
B
B
B
3
4
2
+ 3
+ 9
4
1
2
+
+
+ 5
+ 10
3
+
2
3
2
+ 2
2
2
3
+
3
2
+
+
+ 5
+ 10
1
C
C
C
A
:
(2.4)
We an lassify the xed p oints (
;
) satisfying (2.3) by investigating the Jaobian
of (2.4). One of its eigenvalues is always one, while the other eigenvalue varies ab ove
and b elow one in magnitude. In the left plot of Figure 2, we show the stable portion
of the ellipse (2.3) in blak and the unstable part in white.
We an similarly analyze GMRES(2). This iteration will never progress when, in
addition to the stagnation ondition for GMRES(1),
r
0
also satises
r
T
0
AAr
0
= 0.
For the present example, this requirement implies
2
+ 2
+
2
+ 5
+ 6
+ 1 = 0
;
the equation for an oblique parab ola. This urve intersets the ellipse (2.3) at two
points, drawn as dots in the right plot of Figure 2, the only stagnating residuals
(
; ;
1)
T
for GMRES(2). We an analyze their stability as done above for GMRES(1).
The pro jeted map for this iteration,
r
(2)
k
7!
s
(2)
k
+2
, takes the form
s
(2)
k
+2
=
0
B
B
3
2
3
+ 4
+ 9
4
2
3
+ 4
+ 9
1
C
C
A
:
(2.5)
Analyzing the Jaobian for this GMRES(2) map at the pair of xed p oints, we nd one
to b e unstable (shown in blak in the right plot of Figure 2) while the other is stable
(shown in white). This stable xed p oint is an attrator for stagnating residuals.

RESTARTED GMRES
5
k
r
(
m
)
k
k
k
r
0
k
iteration,
k
GMRES(1)
GMRES(2)
0 5 10 15 20 25 30
10
0
10
5
10
10
10
15
Fig. 3
.
Convergene urves for GMRES
(1)
and GMRES
(2)
applied to
(3.1)
with
x
0
=
0
.
We return briey to the initial residual
r
0
= (2
;
4
;
1)
T
. After the rst few itera-
tions, the angle between
r
(2)
k
and the xed vetor steadily onverges to zero at the
rate 0
:
6452
: : :
suggested by the Jaobian's dominant eigenvalue. We onlude with
high ondene that GMRES(2) never onverges for this initial residual. (If one yle
of GMRES(
m
) pro dues a residual parallel to
r
0
, then either
r
(
m
)
m
=
r
0
or
r
(
m
)
m
=
0
.
Thus a residual an't remain xed in the nite (
;
) plane, but still onverge to
0
.)
3. Seond Example.
The matrix
A
in (2.1) is nondiagonalizable, and one might
be tempted to blame its surprising onvergene behavior on this fat. To demonstrate
that nondiagonalizablity is not an essential requirement, we exhibit a diagonalizable
matrix with eigenvalues
f
1
;
2
;
3
g
for whih restarted GMRES also pro dues extreme
behavior. Take
A
=
0
B
1 2
2
0 2 4
0 0 3
1
C
A
;
b
=
0
3
1
1
1
A
;
(3.1)
with
x
0
=
0
. Again, we onstrut the rst few residuals. For GMRES(1),
r
(1)
1
=
0
2
1
0
1
A
;
r
(1)
2
=
0
2
0
0
1
A
;
r
(1)
3
=
0
0
0
0
1
A
;
while GMRES(2) yields
r
(2)
1
=
0
2
1
0
1
A
;
r
(2)
2
=
0
1
0
1
1
A
;
r
(2)
3
=
1
17
0
8
12
8
1
A
;
r
(2)
4
=
1
67
0
12
12
28
1
A
:
Figure 3 illustrates the onvergene urve for thirty iterations, again omputed using
exat arithmeti.

Citations
More filters
Journal ArticleDOI
TL;DR: Many advances in the development of Krylov subspace methods for the iterative solution of linear systems during the last decade and a half are reviewed.
Abstract: Many advances in the development of Krylov subspace methods for the iterative solution of linear systems during the last decade and a half are reviewed. These new developments include different versions of restarted, augmented, deflated, flexible, nested, and inexact methods. Also reviewed are methods specifically tailored to systems with special properties such as special forms of symmetry and those depending on one or more parameters. Copyright © 2006 John Wiley & Sons, Ltd.

408 citations


Cites background from "The Tortoise and the Hare restart G..."

  • ...We stress however that enlarging the subspace dimension does not always ensure faster convergence; see Eiermann, Ernst, and Schneider [94], Embree [102], for some critical examples and for pointers to further numerical evidence....

    [...]

Journal ArticleDOI
TL;DR: A new technique is presented for accelerating the convergence of restarted GMRES by disrupting this alternating pattern of residual vectors, which resembles a full conjugate gradient method with polynomial preconditioning.
Abstract: We have observed that the residual vectors at the end of each restart cycle of restarted GMRES often alternate direction in a cyclic fashion, thereby slowing convergence. We present a new technique for accelerating the convergence of restarted GMRES by disrupting this alternating pattern. The new algorithm resembles a full conjugate gradient method with polynomial preconditioning, and its implementation requires minimal changes to the standard restarted GMRES algorithm.

194 citations


Additional excerpts

  • ..., [17, 13])....

    [...]

Journal ArticleDOI
TL;DR: The approach presented here can apply not only to conventional processors but also to exotic technologies such as Field Programmable Gate Arrays (FPGA), Graphical Processing Units (GPU), and the Cell BE processor.
Abstract: By using a combination of 32-bit and 64-bit floating point arithmetic, the performance of many sparse linear algebra algorithms can be significantly enhanced while maintaining the 64-bit accuracy of the resulting solution. These ideas can be applied to sparse multifrontal and supernodal direct techniques and sparse iterative techniques such as Krylov subspace methods. The approach presented here can apply not only to conventional processors but also to exotic technologies such as Field Programmable Gate Arrays (FPGA), Graphical Processing Units (GPU), and the Cell BE processor.

100 citations


Cites background from "The Tortoise and the Hare restart G..."

  • ...Assuming for example that the bigger m the better does not guarantee better execution time, and sometimes the convergence can get even worse [20]....

    [...]

  • ...Assuming, for example, that the bigger m the better does not guarantee better execution time, and sometimes the conver­gence canget even worse[Embree2003]....

    [...]

Journal ArticleDOI
TL;DR: This work studies the use of Krylov subspace recycling for the solution of a sequence of slowly-changing families of linear systems, where each family consists of shifted linear systems that differ in the coefficient matrix only by multiples of the identity.

70 citations


Cites methods from "The Tortoise and the Hare restart G..."

  • ...families of shifted linear systems, but it is based on restarted GMRES, meaning the method inherits the properties of stagnation and unpredictable convergence exhibited by restarted GMRES, see, e.g., [12, 17, 29]. An attractive idea would be to combine Frommer and Gl¨assner’s method with a subspace augmentation or deflation technology. Morgan’s GMRES-DR [22] is one candidate, and in [8], this method has been e...

    [...]

Journal ArticleDOI
TL;DR: A simple strategy for varying the restart parameter is proposed and some heuristic explanations for its effectiveness are provided based on analysis of the symmetric case.

61 citations

References
More filters
Journal ArticleDOI
TL;DR: An iterative method for solving linear systems, which has the property of minimizing at every step the norm of the residual vector over a Krylov subspace.
Abstract: We present an iterative method for solving linear systems, which has the property of minimizing at every step the norm of the residual vector over a Krylov subspace. The algorithm is derived from t...

10,907 citations


"The Tortoise and the Hare restart G..." refers background or methods in this paper

  • ...GMRES is an iterative method for solving large nonsymmetric systems of linear equations, Ax = b [8]....

    [...]

  • ...The global optimality of the original algorithm is lost, so although the residual norms remain monotonic, the restarted process can stagnate with a nonzero residual, failing to ever converge [8]....

    [...]

Journal ArticleDOI
TL;DR: The iterative scheme is shown to be a truncation of the standard implicitly shifted QR-iteration for dense problems and it avoids the need to explicitly restart the Arnoldi sequence.
Abstract: The Arnoldi process is a well-known technique for approximating a few eigenvalues and corresponding eigenvectors of a general square matrix. Numerical difficulties such as loss of orthogonality and assessment of the numerical quality of the approximations, as well as a potential for unbounded growth in storage, have limited the applicability of the method. These issues are addressed by fixing the number of steps in the Arnoldi process at a prescribed value k and then treating the residual vector as a function of the initial Arnoldi vector. This starting vector is then updated through an iterative scheme that is designed to force convergence of the residual to zero. The iterative scheme is shown to be a truncation of the standard implicitly shifted QR-iteration for dense problems and it avoids the need to explicitly restart the Arnoldi sequence. The main emphasis of this paper is on the derivation and analysis of this scheme. However, there are obvious ways to exploit parallelism through the matrix-vector ...

1,146 citations


"The Tortoise and the Hare restart G..." refers methods in this paper

  • ...Such effects might also arise from automatic shift-selection strategies in the restarted Arnoldi algorithm for calculating eigenvalues [10]....

    [...]

Journal ArticleDOI
TL;DR: A survey of computational methods in linear algebra can be found in this article, where the authors discuss the means and methods of estimating the quality of numerical solution of computational problems, the generalized inverse of a matrix, the solution of systems with rectangular and poorly conditioned matrices, and more traditional questions such as algebraic eigenvalue problems and systems with a square matrix.
Abstract: The authors' survey paper is devoted to the present state of computational methods in linear algebra. Questions discussed are the means and methods of estimating the quality of numerical solution of computational problems, the generalized inverse of a matrix, the solution of systems with rectangular and poorly conditioned matrices, the inverse eigenvalue problem, and more traditional questions such as algebraic eigenvalue problems and the solution of systems with a square matrix (by direct and iterative methods).

667 citations

Journal Article
TL;DR: In this paper, the authors generalize the Bi-CGSTAB algorithm further, and overcome some shortcomings of BiCGStab2 by combining GMRES(l) and BiCG and profits from both.
Abstract: For a number of linear systems of equations arising from realistic problems, using the Bi-CGSTAB algorithm of van der Vorst [17] to solve these equations is very attractive. Unfortunately, for a large class of equations, where, for instance, Bi-CG performs well, the convergence of BiCGSTAB stagnates. This was observed specifically in case of discretized advection dominated PDE’s. The stagnation is due to the fact that for this type of equations the matrix has almost pure imaginary eigenvalues. With his BiCGStab2 algorithm Gutknecht [5] attempted to avoid this stagnation. Here, we generalize the Bi-CGSTAB algorithm further, and overcome some shortcomings of BiCGStab2. In some sense, the new algorithm combines GMRES(l) and Bi-CG and profits from both.

566 citations


"The Tortoise and the Hare restart G..." refers background in this paper

  • ...One also wonders if related algorithms, including GMRES restarted with an augmented subspace [2] and BiCGSTAB( ) [9], exhibit similarly unusual behavior....

    [...]

Frequently Asked Questions (2)
Q1. What is the olor of a GMRES?

The olor indi ates kr(m)100 k=kr0k on a loga-rithmi s ale: blue regions orrespond to initial residuals that onverge satisfa torily, while the redregions show residuals that stagnate or onverge very slowly. 

For the onve tion-di usion dis retization des ribed in [4℄,GMRES(1) or GMRES(5) an outperform GMRES(20) on moderately re ned grids.