scispace - formally typeset
Search or ask a question
Journal ArticleDOI

J-Orthogonal Matrices: Properties and Generation

01 Jan 2003-Siam Review (Society for Industrial and Applied Mathematics)-Vol. 45, Iss: 3, pp 504-519
TL;DR: This work defines and explores the properties of the exchange operator, which maps J-orthogonal matrices to orthogonalMatrices and vice versa, and shows how the exchange operators can be used to obtain a hyperbolic CS decomposition of a J- Orthogonal matrix directly from the usual CS decompositions of an orthogsonal matrix.
Abstract: A real, square matrix Q is J-orthogonal if QTJQ = J, where the signature matrix $J = \diag(\pm 1)$. J-orthogonal matrices arise in the analysis and numerical solution of various matrix problems involving indefinite inner products, including, in particular, the downdating of Cholesky factorizations. We present techniques and tools useful in the analysis, application, and construction of these matrices, giving a self-contained treatment that provides new insights. First, we define and explore the properties of the exchange operator, which maps J-orthogonal matrices to orthogonal matrices and vice versa. Then we show how the exchange operator can be used to obtain a hyperbolic CS decomposition of a J-orthogonal matrix directly from the usual CS decomposition of an orthogonal matrix. We employ the decomposition to derive an algorithm for constructing random J-orthogonal matrices with specified norm and condition number. We also give a short proof of the fact that J-orthogonal matrices are optimally scaled und...

Content maybe subject to copyright    Report

J -Orthogonal Matrices: Properties and
Generation
Higham, Nicholas J.
2003
MIMS EPrint: 2006.69
Manchester Institute for Mathematical Sciences
School of Mathematics
The University of Manchester
Reports available from: http://eprints.maths.manchester.ac.uk/
And by contacting: The MIMS Secretary
School of Mathematics
The University of Manchester
Manchester, M13 9PL, UK
ISSN 1749-9097

J-ORTHOGONAL MATRICES: PROPERTIES AND GENERATION
NICHOLAS J. HIGHAM
Abstract. A real, square matrix Q is J-orthogonal if Q
T
JQ = J, where the signature matrix
J = diag(±1). J-orthogonal matrices arise in the analysis and numerical solution of various matrix
problems involving indefinite inner products, including, in particular, the downdating of Cholesky
factorizations. We present techniques and tools useful in the analysis, application and construction
of these matrices, giving a self-contained treatment that provides new insights. First, we define and
explore the properties of the exchange operator, which maps J-orthogonal matrices to orthogonal
matrices and vice versa. Then we show how the exchange operator can be used to obtain a hyperbolic
CS decomposition of a J-orthogonal matrix directly from the usual CS decomposition of an orthogonal
matrix. We employ the decomposition to derive an algorithm for constructing random J-orthogonal
matrices with specified norm and condition numb er. We also give a short proof of the fact that
J-orthogonal matrices are optimally scaled under two-sided diagonal scalings. We introduce the
indefinite polar decomposition and investigate two iterations for computing the J-orthogonal polar
factor: a Newton iteration involving only matrix inversion and a Schulz iteration involving only
matrix multiplication. We show that these iterations can be used to J-orthogonalize a matrix that
is not too far from being J-orthogonal.
Key words. J-orthogonal matrix, exchange operator, gyration operator, sweep operator, princi-
pal pivot transform, hyperbolic CS decomposition, two-sided scaling, indefinite least squares problem,
hyperbolic QR factorization, indefinite polar decomposition, Newton’s method, Schulz iteration
AMS subject classifications. 65F30, 15A18
1. Introduction. A matrix Q R
n×n
is J-orthogonal if
Q
T
JQ = J,(1.1)
where J = diag(±1) is a signature matrix. Clearly, Q is nonsingular and QJQ
T
= J.
This type of matrix arises in hyperbolic problems, that is, problems where there is
an underlying indefinite inner product or weight matrix. We give two examples to
illustrate the utility of J-orthogonal matrices.
First consider the downdating problem of computing the Cholesky factorization of
a positive definite matrix C = A
T
A B
T
B, where A R
p×n
(p n) and B R
q ×n
.
This task arises when solving a regression problem after some of the rows (namely
those of B) of the data matrix are removed, and A in this case is usually upper
triangular. Numerical stability considerations dictate that we should avoid explicit
formulation of C. If we can find a J-orthogonal matrix Q such that
Q
A
B
=
R
0
,(1.2)
with J = diag(I
p
, I
q
) and R R
n×n
upper triangular, then
C =
A
B
T
J
A
B
=
A
B
T
Q
T
JQ
A
B
= R
T
R,
Numerical Analysis Report 408, Manchester Centre for Computational M athematics, September
2002. Revised March 2003.
Department of Mathematics, University of Manchester, Manchester, M13 9PL, England
(higham@ma.man.ac.uk, http://www.ma.man.ac.uk/~higham/). This work was supported by Engi-
neering and Physical Sciences Research Council grant GR/R22612.
1

2 NICHOLAS J. HIGHAM
so R is the desired Cholesky factor. The factorization (1.2) is a hyperbolic QR fac-
torization; for details of how to compute it see, for example, [1].
A second example where J-orthogonal matrices play a key role is in the solution
of the symmetric definite generalized eigenproblem Ax = λBx, where A and B are
symmetric, some linear combination of them is positive definite, and B is nonsingular.
Through the use of a congruence transformation (for example by using a block LDL
T
decomposition of B followed by a diagonalization of the block diagonal factor [38]) the
problem can be reduced to
e
Ax = λJx, for some signature matrix J = diag (±1). If we
can find a J-orthogonal Q such that Q
T
e
AQ = D = diag (d
i
) then the eigenvalues are
the diagonal elements of JD; such a Q can be constructed using a Jacobi algorithm
of Veseli´c [40].
In addition to these practical applications, J-orthogonal matrices are of significant
theoretical interest. For example, they play a fundamental role in the study of J-
contractive matrices [30], which are matrices X for which XJX
T
J, where A 0
denotes that the symmetric matrix A is p ositive semidefinite.
A matrix Q R
n×n
is (J
1
, J
2
)-orthogonal if
Q
T
J
1
Q = J
2
,(1.3)
where J
1
= diag (±1) and J
2
= diag (±1) are signature matrices having the same
inertia. (J
1
, J
2
)-orthogonal matrices are also known as hyperexchange matrices and
J-orthogonal matrices as hypernormal matrices [2]. Since J
1
and J
2
in (1.3) have the
same inertia, J
2
= P J
1
P
T
for some permutation matrix P , and hence (QP )
T
J
1
(QP ) =
J
1
. A (J
1
, J
2
)-orthogonal matrix is therefore simply a column permutation of a J
1
-
orthogonal matrix, and so for the purposes of this work we can restrict our attention
to J-orthogonal matrices. An application in which (J
1
, J
2
)-orthogonal matrices arise
with J
1
and J
2
generally different is the HR algorithm of Brebner and Grad [5] and
Bunse-Gerstner [6] for solving the standard eigenvalue problem for J-symmetric ma-
trices. A matrix A R
n×n
is J-symmetric if AJ is symmetric, or, equivalently, if
JA, JA
T
or A
T
J is symmetric. Given a J
0
-symmetric matrix A, the kth stage of the
unshifted HR algorithm consists of factoring A
k
= H
k
R
k
, where H
k
is (J
k
, J
k+1
)-
orthogonal with J
k+1
= H
T
k
J
k
H
k
and R
k
is upper triangular, and then setting
A
k+1
= R
k
H
k
. Computational details and convergence properties of the algorithm
can be found in [6].
Unlike the subclass of orthogonal matrices, J-orthogonal matrices can be arbi-
trarily ill conditioned. This poses interesting questions and difficulties in the design,
analysis and testing of algorithms and motivates our attempt to gain a better under-
standing of the class of J-orthogonal matrices.
The purpose of this paper is threefold. First we collect some interesting and not
so well-known properties of J-orthogonal matrices. In particular, we give a new proof
of the hyperbolic CS decomposition via the usual CS decomposition by exploiting the
exchange operator. The exchange operator is a tool that has found use in several areas
of mathematics and is known by several different names; we give a brief survey of its
properties and its history. We also give a new proof of the fact that J-orthogonal
matrices are optimally scaled under two-sided diagonal scalings. Our second aim is to
show how to generate random J-orthogonal matrices with specified singular values,
and in particular with specified norms and condition numbers—a capability that is
very useful for constructing test data for problems with an indefinite flavour. Finally,
we investigate two Newton iterations for computing a J-orthogonal matrix, one involv-
ing only matrix inversion, the other only matrix multiplication. Both iterations are

J-ORTHOGONAL MATRICES 3
shown to conver ge to the J-orthogonal factor in a certain indefinite polar decomposi-
tion under suitable conditions. Analogously to the case of orthogonal matrices and the
corresponding Newton iterations [15], [20], we show that these Newton iterations can
be used to J-orthogonalize a matrix that is not too far from being J-orthogonal. An
application is to the situation where a matrix that should be J-orthogonal turns out
not to be because of rounding or other errors and it is desired to J-orthogonalize it.
J-orthogonal matrices, and hyperbolic problems in general, are the subject of
much recent and current research, covering both theory and algorithms. This paper
provides a self-contained treatment that highlights some techniques and tools useful
in the analysis and application of these matrices; the treatment should also be of more
general interest.
Throughout, we take J to have the form
J =
I
p
0
0 I
q
, p + q = n,(1.4)
and we use exclusively the 2-norm: kAk
2
= max
x6=0
kAxk
2
/kxk
2
, where kxk
2
2
= x
T
x.
2. The exchange operator. Let A R
n×n
and consider the system
y =
1
p y
1
q y
2
=
p q
p A
11
A
12
q A
21
A
22
1
x
1
x
2
p
q
= Ax,(2.1)
where A
11
is nonsingular. We use this partitioning of A throughout the section. By
solving the first equation in (2.1) for x
1
and then eliminating x
1
from the second
equation we obtain
x
1
y
2
= exc(A)
y
1
x
2
,(2.2)
where
exc(A) =
A
1
11
A
1
11
A
12
A
21
A
1
11
A
22
A
21
A
1
11
A
12
.
We call exc the exchange operator, since it exchanges x
1
and y
1
in (2.1). Note that
the (2,2)-block of exc(A) is the Schur complement of A
11
in A. The definition of
the exchange operator can be generalized to allow the “pivot matrix” A
11
to be any
principal submatrix, but for our purposes this extra level of generality is not necessary.
It is easy to see that the exchange operator is involutary,
exc(exc(A)) = A,(2.3)
and moreover that
exc(JAJ) = Jexc(A)J = exc(A
T
)
T
.(2.4)
This last identity shows that J is naturally associated with exc.
We first address the nonsingularity of exc(A). The blo ck LU factorization
exc(A) =
I 0
A
21
A
22
A
1
11
A
1
11
A
12
0 I
=
I 0
A
21
A
22
A
11
A
12
0 I
1
LR
1
(2.5)

4 NICHOLAS J. HIGHAM
will be useful.
Lemma 2.1. Let A R
n×n
with A
11
nonsingular. Then exc(A) is nonsingular if
and only if A
22
is nonsingular. If A is nonsingular and exc(A
1
) exists then exc(A)
is nonsingular and
exc(A)
1
= exc(A
1
).(2.6)
Proof. For A R
n×n
with A
11
nonsingular, the block LU factorization (2.5)
makes clear that exc(A) is nonsingular if and only if A
22
is nonsingular. The last part
is obtained by rewriting (2.1) as x = A
1
y and deriving the corresponding analogue
of (2.2):
y
1
x
2
= exc(A
1
)
x
1
y
2
.(2.7)
It follows from (2.7) that for any x
1
and y
2
there is a unique x
2
and y
1
, which implies
from (2.2) that exc(A) is nonsingular and exc(A)
1
= exc(A
1
).
Note that either of A and exc(A) can be singular without the other being singular,
as shown by the examples with p = q = 1,
A =
1 1
1 0
, exc(A) =
1 1
1 1
, A =
1 1
1 1
, exc(A) =
1 1
1 0
.
For completeness, we mention that for both exc(A) and exc(A
1
) to exist and be
nonsingular, it is necessary and sufficient that A, A
11
and A
22
be nonsingular.
The reason for our interest in the exchange operator is that it maps J-orthogonal
matrices to orthogonal matrices and vice versa. Note that J-orthogonality of A implies
that A
T
11
A
11
= I + A
T
21
A
21
and hence that A
11
is nonsingular and exc(A) exists, but
if A is orthogonal A
11
can be singular.
Theorem 2.2. Let A R
n×n
. If A is J-orthogonal then exc(A) is orthogonal.
If A is orthogonal and A
11
is nonsingular then exc(A) is J-orthogonal.
Proof. Proving the result by working directly with exc(A) involves some laborious
algebra. A more elegant proof involving quadratic forms is given by Stewart and
Stewart [36, sec. 2]. We give another proof, suggested by Chris Paige. Assume first
that A is orthogonal with A
11
nonsingular. Then exc(A
T
) = exc(A
1
) exists and
Lemma 2.1 shows that exc(A) is nonsingular and exc(A)
1
= exc(A
1
) = exc(A
T
).
Hence, using (2.4),
I = exc(A
T
)exc(A) = Jexc(A)
T
J · exc(A),
which shows that exc(A) is J-orthogonal.
If A is J-orthogonal then, as noted above, A
11
is nonsingular. Also JA
T
J = A
1
and so from Lemma 2.1, exc(JA
T
J) = exc(A
1
) = exc(A)
1
. But (2.4) shows that
exc(JA
T
J) = exc(A)
T
, and we conclude that exc(A) is orthogonal.
As an example of a result of a different flavour, we give the following generalization
to arbitrary p of a result obtained by Duffin, Hazony, and Morrison [10] for p = 1.
Theorem 2.3. Let A R
n×n
with A
11
nonsingular. Then exc(A) + exc(A)
T
is
congruent to A + A
T
.
Proof. Using (2.5) we have
exc(A) + exc(A)
T
= LR
1
+ R
T
L
T
= R
T
(R
T
L + L
T
R)R
1
.

Citations
More filters
Journal ArticleDOI
TL;DR: A key feature of this analysis is the identification of two particular classes of scalar products, termed unitary and orthosymmetric, which serve to unify assumptions for the existence of structured factorizations.
Abstract: Let $A$ belong to an automorphism group, Lie algebra, or Jordan algebra of a scalar product. When $A$ is factored, to what extent do the factors inherit structure from $A$? We answer this question for the principal matrix square root, the matrix sign decomposition, and the polar decomposition. For general $A$, we give a simple derivation and characterization of a particular generalized polar decomposition, and we relate it to other such decompositions in the literature. Finally, we study eigendecompositions and structured singular value decompositions, considering in particular the structure in eigenvalues, eigenvectors, and singular values that persists across a wide range of scalar products. A key feature of our analysis is the identification of two particular classes of scalar products, termed unitary and orthosymmetric, which serve to unify assumptions for the existence of structured factorizations. A variety of different characterizations of these scalar product classes are given.

83 citations

Journal ArticleDOI
TL;DR: In this paper, a generalized tensor function according to the tensor singular value decomposition (T-SVD) is defined, from which the projection operators and Moore-Penrose inverse of tensors are obtained.

81 citations

Journal ArticleDOI
TL;DR: It is shown that group structure is preserved precisely when $f(A^{-1}) = f(A)^{- 1}$ for bilinear forms and when $F(A)\in\mathbb{G}$ is the matrix automorphism group associated with a bil inear or sesquilinear form, and meromorphic functions that satisfy each of these conditions are characterized.
Abstract: For which functions $f$ does $A\in\mathbb{G} \Rightarrow f(A)\in\mathbb{G}$ when $\mathbb{G}$ is the matrix automorphism group associated with a bilinear or sesquilinear form? For example, if $A$ is symplectic when is $f(A)$ symplectic? We show that group structure is preserved precisely when $f(A^{-1}) = f(A)^{-1}$ for bilinear forms and when $f(A^{-*}) = f(A)^{-*}$ for sesquilinear forms. Meromorphic functions that satisfy each of these conditions are characterized. Related to structure preservation is the condition $f(\overline{A}) = \overline{f(A)}$, and analytic functions and rational functions satisfying this condition are also characterized. These results enable us to characterize all meromorphic functions that map every $\mathbb{G}$ into itself as the ratio of a polynomial and its ``reversal,'' up to a monomial factor and conjugation. The principal square root is an important example of a function that preserves every automorphism group $\mathbb{G}$. By exploiting the matrix sign function, a new family of coupled iterations for the matrix square root is derived. Some of these iterations preserve every $\mathbb{G}$; all of them are shown, via a novel Frechet derivative-based analysis, to be numerically stable. A rewritten form of Newton's method for the square root of $A\in\mathbb{G}$ is also derived. Unlike the original method, this new form has good numerical stability properties, and we argue that it is the iterative method of choice for computing $A^{1/2}$ when $A\in\mathbb{G}$. Our tools include a formula for the sign of a certain block $2\times 2$ matrix, the generalized polar decomposition along with a wide class of iterations for computing it, and a connection between the generalized polar decomposition of $I+A$ and the square root of $A\in\mathbb{G}$.

68 citations

Journal ArticleDOI
TL;DR: In this paper, an extensive and unified collection of structure-preserving transformations for non-degenerate bilinear or sesquilinear forms on R n or C n is presented.
Abstract: An extensive and unified collection of structure-preserving transformations is pre- sented and organized for easy reference. The structures involved arise in the context of a non- degenerate bilinear or sesquilinear form on R n or C n . A variety of transformations belonging to the automorphism groups of these forms, that imitate the action of Givens rotations, Householder reflec- tors, and Gauss transformations are constructed. Transformations for performing structured scaling actions are also described. The matrix groups considered in this paper are the complex orthogonal, real, complex and conjugate symplectic, real perplectic, real and complex pseudo-orthogonal, and pseudo-unitary groups. In addition to deriving new transformations, this paper collects and unifies existing structure-preserving tools.

68 citations

01 Jan 2011
TL;DR: In this paper, a model for simultaneous diagonalization and damping of J-Hermitian matrices is presented, and a spectral decomposition of a general Hermitian matrix is presented.
Abstract: 1 The model.- 2 Simultaneous diagonalisation (Modal damping).- 3 Phase space.- 4 The singular mass case.- 5 "Indefinite metric".- 6 Matrices and indefinite scalar products.- 7 Oblique projections.- 8 J-orthogonal projections.- 9 Spectral properties and reduction of J-Hermitian matrices.- 10 Definite spectra.- 11 General Hermitian matrix pairs.- 12 Spectral decomposition of a general J-Hermitian matrix.- 13 The matrix exponential.- 14 The quadratic eigenvalue problem.- 15 Simple eigenvalue inclusions.- 16 Spectral shift.- 17 Resonances and resolvents.- 18 Well-posedness .- 19 Modal approximation.- 20 Modal approximation and overdampedness.- 21 Passive control.- 22 Perturbing matrix exponential.- 23 Notes and remarks.

65 citations

References
More filters
Book
01 Jan 1985
TL;DR: In this article, the authors present results of both classic and recent matrix analyses using canonical forms as a unifying theme, and demonstrate their importance in a variety of applications, such as linear algebra and matrix theory.
Abstract: Linear algebra and matrix theory are fundamental tools in mathematical and physical science, as well as fertile fields for research. This new edition of the acclaimed text presents results of both classic and recent matrix analyses using canonical forms as a unifying theme, and demonstrates their importance in a variety of applications. The authors have thoroughly revised, updated, and expanded on the first edition. The book opens with an extended summary of useful concepts and facts and includes numerous new topics and features, such as: - New sections on the singular value and CS decompositions - New applications of the Jordan canonical form - A new section on the Weyr canonical form - Expanded treatments of inverse problems and of block matrices - A central role for the Von Neumann trace theorem - A new appendix with a modern list of canonical forms for a pair of Hermitian matrices and for a symmetric-skew symmetric pair - Expanded index with more than 3,500 entries for easy reference - More than 1,100 problems and exercises, many with hints, to reinforce understanding and develop auxiliary themes such as finite-dimensional quantum systems, the compound and adjugate matrices, and the Loewner ellipsoid - A new appendix provides a collection of problem-solving hints.

23,986 citations

Book
01 Jan 1990
TL;DR: In this article, the Perturbation of Eigenvalues and Generalized Eigenvalue Problems are studied. But they focus on linear systems and Least Squares problems and do not consider invariant subspaces.
Abstract: Preliminaries. Norms and Metrics. Linear Systems and Least Squares Problems. The Perturbation of Eigenvalues. Invariant Subspaces. Generalized Eigenvalue Problems.

2,754 citations

01 Jan 2002
TL;DR: Higham as discussed by the authors gives a thorough, up-to-date treatment of the behavior of numerical algorithms in finite precision arithmetic, combining algorithmic derivations, perturbation theory, and rounding error analysis.
Abstract: From the Publisher: What is the most accurate way to sum floating point numbers? What are the advantages of IEEE arithmetic? How accurate is Gaussian elimination and what were the key breakthroughs in the development of error analysis for the method? The answers to these and many related questions are included here. This book gives a thorough, up-to-date treatment of the behavior of numerical algorithms in finite precision arithmetic. It combines algorithmic derivations, perturbation theory, and rounding error analysis. Software practicalities are emphasized throughout, with particular reference to LAPACK and MATLAB. The best available error bounds, some of them new, are presented in a unified format with a minimum of jargon. Because of its central role in revealing problem sensitivity and providing error bounds, perturbation theory is treated in detail. Historical perspective and insight are given, with particular reference to the fundamental work of Wilkinson and Turing, and the many quotations provide further information in an accessible format. The book is unique in that algorithmic developments and motivations are given succinctly and implementation details minimized, so that attention can be concentrated on accuracy and stability results. Here, in one place and in a unified notation, is error analysis for most of the standard algorithms in matrix computations. Not since Wilkinson's Rounding Errors in Algebraic Processes (1963) and The Algebraic Eigenvalue Problem (1965) has any volume treated this subject in such depth. A number of topics are treated that are not usually covered in numerical analysis textbooks, including floating point summation, block LU factorization, condition number estimation, the Sylvester equation, powers of matrices, finite precision behavior of stationary iterative methods, Vandermonde systems, and fast matrix multiplication. Although not designed specifically as a textbook, this volume is a suitable reference for an advanced course, and could be used by instructors at all levels as a supplementary text from which to draw examples, historical perspective, statements of results, and exercises (many of which have never before appeared in textbooks). The book is designed to be a comprehensive reference and its bibliography contains more than 1100 references from the research literature. Audience Specialists in numerical analysis as well as computational scientists and engineers concerned about the accuracy of their results will benefit from this book. Much of the book can be understood with only a basic grounding in numerical analysis and linear algebra. About the Author Nicholas J. Higham is a Professor of Applied Mathematics at the University of Manchester, England. He is the author of more than 40 publications and is a member of the editorial boards of the SIAM Journal on Matrix Analysis and Applications and the IMA Journal of Numerical Analysis. His book Handbook of Writing for the Mathematical Sciences was published by SIAM in 1993.

1,565 citations

Book
01 Jan 1998
TL;DR: This volume treats the numerical solution of dense and large-scale eigenvalue problems with an emphasis on algorithms and the theoretical background required to understand them.
Abstract: This book is the second volume in a projected five-volume survey of numerical linear algebra and matrix algorithms. This volume treats the numerical solution of dense and large-scale eigenvalue problems with an emphasis on algorithms and the theoretical background required to understand them. Stressing depth over breadth, Professor Stewart treats the derivation and implementation of the more important algorithms in detail. The notes and references sections contain pointers to other methods along with historical comments. The book is divided into two parts: dense eigenproblems and large eigenproblems. The first part gives a full treatment of the widely used QR algorithm, which is then applied to the solution of generalized eigenproblems and the computation of the singular value decomposition. The second part treats Krylov sequence methods such as the Lanczos and Arnoldi algorithms and presents a new treatment of the Jacobi-Davidson method. The volumes in this survey are not intended to be encyclopedic. By treating carefully selected topics in depth, each volume gives the reader the theoretical and practical background to read the research literature and implement or modify new algorithms. The algorithms treated are illustrated by pseudocode that has been tested in MATLAB implementations.

653 citations

Journal ArticleDOI
TL;DR: Applications of the polar decomposition to factor analysis, aerospace computations and optimisation are outlined; and a new method is derived for computing the square root of a symmetric positive definite matrix.
Abstract: A quadratically convergent Newton method for computing the polar decomposition of a full-rank matrix is presented and analysed. Acceleration parameters are introduced so as to enhance the initial rate of convergence and it is shown how reliable estimates of the optimal parameters may be computed in practice.To add to the known best approximation property of the unitary polar factor, the Hermitian polar factor H of a nonsingular Hermitian matrix A is shown to be a good positive definite approximation to Aand $\frac{1}{2}(A + H)$ is shown to be a best Hermitian positive semi-definite approximation to A. Perturbation bounds for the polar factors are derived.Applications of the polar decomposition to factor analysis, aerospace computations and optimisation are outlined; and a new method is derived for computing the square root of a symmetric positive definite matrix.

446 citations


"J-Orthogonal Matrices: Properties a..." refers background or methods in this paper

  • ..., [15], [23]), the latter iteration being produced on setting J = I....

    [...]

  • ...Analogously to the case of orthogonal matrices and the corresponding Newton iterations [15], [20], we show that these Newton iterations can be used to J-orthogonalize a matrix that is not too far from being J-orthogonal....

    [...]

Frequently Asked Questions (5)
Q1. What are the contributions in "J-orthogonal matrices: properties and generation higham, nicholas j. 2003" ?

The authors present techniques and tools useful in the analysis, application and construction of these matrices, giving a self-contained treatment that provides new insights. Then the authors show how the exchange operator can be used to obtain a hyperbolic CS decomposition of a J-orthogonal matrix directly from the usual CS decomposition of an orthogonal matrix. The authors introduce the indefinite polar decomposition and investigate two iterations for computing the J-orthogonal polar factor: a Newton iteration involving only matrix inversion and a Schulz iteration involving only matrix multiplication. The authors show that these iterations can be used to J-orthogonalize a matrix that is not too far from being J-orthogonal. 

Restoring lost orthogonality is a common requirement, for example in numerical solution of matrix differential equations having an orthogonal solution [17], or for computed eigenvector matrices of symmetric matrices. 

From standard analysis of this iteration (see, e.g., [23]) the authors know that Sk converges quadratically to sign(S0), which is the identity matrix since the spectrum of S0 lies in the open right half-plane. 

Unlike for orthogonal matrices, for general J-orthogonal matrices ‖Q‖2 can be arbitrarily large and this has implications for the attainable accuracy of the Newton and Schulz iterations in floating point arithmetic. 

Such an iteration can be obtained by adapting the Schulz iteration, which exists in variants for computing the matrix inverse [31], the orthogonal polar factor [20], the matrix sign function [22], and the matrix square root [18].