Numerische Mathematik: K. Veseli
Numerische Mathematik: K. Veseli
yields diagonal matrices S', T'. We consider here the most important class of pairs
S, T which are simultaneously diagonalizable, namely the class of definite matrix
242 K. Veseli6
(1.4) T = W J W x, J =
0]
-I,_,, "
Note that in this way the initial symmetry of the problem is preserved during
the whole diagonalizing process. If a definitizing shift # is known in advance then
the step (1.5) can be made with the positive definite matrix T ' = T - # S thus
reducing to a common Cholesky decomposition with J = I. The eigenreduction
may then be continued by diagonalizing a single symmetric matrix A. However, the
transition from S, T to `4, J may be unstable even if T itself is positive definite (see
" [15], Ex. 8.7.2). There are pairs where this transition is safe ( [2, 39] ) and this issue
deserves further research. A possible symmetry preserving method for a general
symmetric definite pair is the Jacobi-like algorithm of Falk and Langemeyer [13]
for which the global convergence has been proved in [19] whereas [32] proves the
quadratic convergence for distinct eigenvalues. With multiple eigenvalues this
method usually lacks the quadratic convergence [20]. This calls for a further
investigation. In this paper we consider cases in which the reduction (1.5) is
relatively simple. We shall present two main applications of our Jacobi method:
updated in the factorized form (this is the "one-sided" or "implicit" variant of the
Jacobi method) so that, upon convergence, the factor W contains the eigenvectors
of T. This seems rather a roundabout way to diagonalize a real symmetric matrix
and uses hyperbolic transformations besides! Nevertheless, this algorithm is appar-
ently the closest of all known algorithms to the ideal: if the matrix is "well-behaved"
i.e. if small relative errors in the matrix entries cause small relative errors in the
eioenvalues then compute the latter with nearly the same accuracy. For positive
definite T we have J = I and our method reduces to a known one [37] which is
shown to attain the above ideal [7]. For indefinite T we still obtain a positive
definite current matrix A whereas the initial indefiniteness is "packed" in J which
does not change during the Jacobi process (note that indefinite J means the use of
hyperbolic transformations). Pairs A, J with A positive definite allow a very sharp
"relative" perturbation theory [38]. This is a basis for an error analysis, performed
in the forthcoming dissertation of Slapni6ar [33] whose results are a non-trivial
generalization of those in [7]. As the Slapni~ar's analysis shows, the orthogonality
of the transformation plays only a secondary role in the error estimates. Our
method is generally more accurate than the standard ql or Jacobi method, applied
directly to the initial matrix T. Our method is also generally faster than the
standard Jacobi method. This is due to the factorized form of the current matrix as
well as to the fact that the decomposition (1.4) is a preconditioner reducing the
number of the Jacobi iterations. The rounding errors in performing (1.4) are, as
a rule, more serious than those in the Jacobi method that follows ([33]) and it is
certainly of great help if the factor W is accessible without computing T itself.
The matrices M, K, D are real symmetric and positive definite. The pencil (1.6) is
called overdamped [10], if
(1.7) (xTDx) 2 > 4 x T M x x T K x
which is equivalent to (1.6). Note that both matrices in (1.9) are indefinite. Using
the decomposition
1 Recently, Scott and Ward [30] proposed an algorithm for a partial eigenreduction of large
sparse matrices generated by overdamped pencils. Our algorithm can be considered as its
counterpart for small full matrices. It could therefore be used as a subspace eigenreduction routine
within the mentioned algorithm of Scott and Ward.
A Jacobi eigenreduction algorithm for definite matrix pairs 245
(the non-displayed elements are those of the unit matrix) having the property
(2.3) CT J C = J .
([cos x -sinx 1
(2.4) F c" c,s~=,)l_sinx cosx_l' l<r,s<m or m + l < r , s < n ,
kC~ c~s_J )Fcoshy sinhy 1
[,Lsinhy coshyJ' otherwise.
or
2ars
(2.6) tanh 2y = t, t=
art + a~s "
In (2.6) of course,
(2.7) larr + a~st > 21a,~],
has to be valid. We call the two types of transformations in (2.4) the trigonometric
and the hyperbolic transformation, respectively. In a real computation we use the
stable Rutishauser formulae (or their hyperbolic analogs) from 1-41]).
If (2.7) is violated at some step, then the algorithm breaks down.
Theorem 2.1. Let A' be given by (2.1)-(2.3). Then the pair A', J is positive (negative)
definite i f and only if A, J is such. In this case we have f o r r < m < s
(2.8)
I--IJ
2a,s
a,, + a~
<
=
1
(ar, + a ~ ) 2 "
246 K. Veseli6
If A - / ~ J is, say, positive definite and r < m < s then the matrix
a,~ ass + #
(2.10)
F
A = LAT2 A2~ '
F
= LA~ Ai:J'
where the partition is that of d in (1.5), the invariance properties (2.9) can be
expressed as
(2.11) TrA'I 1 - TrA ~2 = TrA 1i - TrA 2z ,
(2.12) l[a]l[I 2 + Ila~2ll z - 211A'12112 = Ilaixll 2 + 11A22112 -2llA12[I 2
A Jacobi eigenreduction algorithm for definite matrix pairs 247
$2(A) = Z a2
i,j
iee j
for a general matrix A. Subtracting (2.15) from (2.12) and using the identity
lIAr2 [] = IIA 12 II ,
(2.20)
~
/' = tanh 2y = It,
t ma~sign t, [t] > tmax
otherwise ,
Lemma 2.2. Let A, J be, say, positive definite. Perform a hyperbolic step with any
parameter 33. Then
(2.21) TrA - TrA' = (a,, + as~)[1 - cosh 233 + t sinh 233],
248 K. Veseli6
(t from (2.6)). The above function is strictly concave in ~ and takes its maximum at
29 = artanh t. With 33 = artanh ?, ?from (2.20)
where A 1 is the initial matrix, 3 is the length of the interval J in (1.2). In particular,
the choice l?I = t tm,x ( can occur only finitely many times.
Proof The formulae (2.21) and (2.22) are directly verified. From (2.22) and the
invariance of the trace under trigonometric transformations it follows that TrA is
non-increasing. It is obviously bounded from below by #TrJ, p any definitizing
shift, and has therefore a limit tro. Now (2.23) follows from a r t + a~ > 3
(Th. 2.1). Q.E.D.
Note that for the validity of the identities (2.21), (2.22) no definiteness of the pair
A, J is necessary. It is sufficient that an r, s-hyperbolic transformation is possible i.e.
that (2.7) holds and that art + as~ is, say, positive.
The estimate (2.23) gives an upper bound to the condition of the hyperbolic
transformation independently of the choice of tm,~. For matrices near the bound-
ary of the definiteness region, however, this bound is of no use and a convenient
t~,,. should be introduced. It is plausible that tmax should be as close to 1 as
possible while still producing a tolerable condition of the hyperbolic matrix C in
(2.4). Our experiments suggested t max = 0.8 as a good value (see Sect. 4 below). In
this ease
Theorem 2.3. Let A, J be definite. Then the Jacobi algorithm with the row-cyclic
strategy is convergent i.e. Ak converges to a diagonal matrix as k -* oo.
Proof Let
A1 = A , A2 = C ~ A 1 C 1 . . . .
A Jacobi eigenreduction algorithm for definite matrix pairs 249
be the sequence generated by the Jacobi algorithm. For y sufficiently smalt we have
tan~=tany, -~ <~< ~,
Another important class of strategies uses thresholds i.e. in sweeping the matrix
in a cyclic order a rotation is performed, if the pivot off-diagonal is larger than
a given threshold. The simplest way is just to take one, small enough, threshold and
use it as the stopping criterion as well. The following convergence theorem will
include matrix pairs A, J whose definiteness is not known in advance. We will show
that the method either converges or detects the indefiniteness in finitely many steps.
For this purpose we shall introduce three criteria simple to check at each step.
The criteria will be shown to hold for any definite matrix pair. We will then show
that the algorithm converges--at least with a threshold strategy - - if they are never
violated. The three criteria are:
Criterion II. For all pivots r, s with r < m < s (hyperbolic transformations) the
quantity
is constant.
250 K. Veseli6
- max aii , (7 ~ - 1 ,
l<_i<m
- max aii, a = 1 ,
O~1 .~_ m < j < n
- max aii , t7 = -- 1
m<j<=n
The current matrix A k has to satisfy the following conditions at every stage of
the process
a=l: a=-l:
~(k)
a~) > c t z , l<i<_m; "u <a2, l<i<_m;
(k) (k)
a~j > c t l , m<j<n; a;j <~1, m<j<n.
It is clear that I and II must be fulfilled for any definite pair A, J (Th. 2.1). N o w
consider III e.g. for a positive definite pair (a = 1). Since both A 1 - g J and
Ak -- # J are positive definite for some ~t we obtain
have to be made on the current matrix A after the transformation on r, s has been
performed.
W e now state our most general convergence theorem.
Theorem 2.4. L e t the Jacobi method follow any cyclic strategy by rotating on all r, s
with
for a given positive threshold a. I f the Criteria 1 - I I 1 are fulfilled then the number o f
steps is finite. I f a sequence a 1, a t . . . . ~ 0 o f such thresholds is given then the Jaeobi
algorithm is convergent.
A Jacobi eigenreduction algorithm for definite matrix pairs 251
P r o o f Let A ~ = A, A 2 . . . .
be the sequence of matrices generated by the Jacobi
algorithm. We use the natural notation
(l~ A(k) 7
[A "a12 | , (k))
Let tr from Criterion II be, say, 1. Again (2.22) (see the remarks following
L e m m a 2.2) implies that TrA]~, TrA ~k~ and T r A k are non-increasing sequences.
According to Criterion III we have
(2.29)
j=l
for m < i < n. Thus, the sequences al~ ) are bounded in k. The same is therefore true
for TrA~ kF, Z r A ,E(k)
2 and T r A k as well. For k large enough (2.22) and (2.25) yield
M = T r A i - ( n - 2 ) m a x ( [ ~ l [ , t~21)
Since the sequence Ak is bounded (2.3t) implies that hyperbolic rotations occur
only finitely m a n y times. The remaining trigonometric transformations are also
finitely m a n y since (2.17) implies S 2 ( A k + l ) < S 2 ( A k ) - - 20" 2. N o w it is clear that
with a sequence a l, a 2 . . . . ~ 0 of thresholds the off-diagonals will tend to zero.
The proof of the convergence of the diagonal elements is again omitted. Q.E.D.
(2.32)
where a is a given threshold with 0 < a < 1 and stops, if (2.32) is false for all r 4= s
and a is small enough. This stopping criterion is useful, if the algorithm is expected
to compute the eigenvalues with uniformly small relative accuracy. Indeed, if
]a,, t < a I x f ~ , l [a,~ I for all r = s then by [2] the eigenvalues lie in the union of the
intervals
Although there is little doubt that our algorithm converges under this strategy as
well (our experiments confirm that) we have a p r o o f only for the important special
case where A itself is positive definite (case d ) .
Theorem 2.5. Let A be positive definite. Then all assertions of Th. 2.4 are true, if the
threshold criterion (2.28) is substituted by (2.32).
Proof. By
detA
X~(A) -
a l l 9 . . ann
denote the H a d a m a r d measure of A. The positive definiteness implies
0 < o~g(A) < 1, where 9f'(A) = 1 holds, if and only if A is diagonal. Since the
transformation (2.1) preserves both the positive definiteness and the determinant
we have
1 - a'
X~(A ') = X~(A) - - ,
1--~x
with
a,2 ars ~2
~X= - - (Xr --
arrass ' arr,as s, "
With tanh y = t from (2.6) we have a' = 0 and
3~(Ak)
3f(Ak+l) > -
1 - - tr
Thus, there are only finitely m a n y rotations for a given threshold. With the
modified parameter (2.20) a straightforward, if lengthy, calculation leads to
1 2
fftma
-- x
3V(Ak+~) > 3V(A,)
(1 - atmax)2 '
with the same conclusion as in Th. 2.4. 2 Q.E.D.
The quadratic convergence of the method has recently been proved in [9] in full
analogy to the standard Jacobi algorithm.
If the m e t h o d is convergent then the value tro from L e m m a 2.2 is given by
n--tit
tro= 2i ~ - ~ 27 9
i=1 i=1
2 The use of the Hadamard measure in proving the convergence ofa Jacobi method goes back to
Gose [16]. Note that this type of proof makes absolutely no difference between trigonometric and
hyperbolic rotations.
A Jacobi eigenreduction algorithm for definite matrix pairs 253
(3.1) A = G J G T, J =
['0 ~ -I
where G has full column rank, the diagonal blocks in d need not have the same
'
is positive definite and our Jacobi method is applied to the pair ~{, J by updating
the factor
(3.2) G1 = G, G k + 1 --- G k C k .
where the diagonal matrix Jy 1/2 carries the non-vanishing eigenvalues of A. The
current matrix A k is not updated, except the diagonals whose updating is non-
expensive (see Rutishauser formulae in [41]). The pivot element d~ ), needed for the
rotation parameter, is computed as the scalar product of the columns of Gk. All
d(k)
,, are refreshed from G k after every sweep. In this way only one square array of
storage suffices to compute both the eigenvalues and the eigenvectors. The
stopping criterion
(3.3) ar~
-~(k) I <
= a ~ / l a ( ~ ) I u~
_~(~) t, for a l l r 4 : s
3 The idea of implicit Jacobi algorithms goes back to Hestenes 1-22] (see also [6, 18, 23, 37]).
, Here we obviously have to decide about the numerical rank of A. This is a delicate question
which deserves a careful treatment. In view of the fact that small eigenvalues may be equally well
defined as the large ones a simple possibility is to accept rank defect only, if during the
decomposition (3.1) exact zeros are encountered (I-33]).
254 K. Veseli6
yields not only a desired relative accuracy of the eigenvalues but also the desired
orthogonality of the eigenvector matrix. 5 In the case J = I (A positive definite) this
method reduces to a known one in 1-37, 7].
Diagonalizing the pair A, J from (1.11). Here A and J have the same inertia; more
precisely
(3.4) A = H J H T, H = LL?IL3 ,
(3.5) A= HTjH,J.
L e m m a 3.1. The pair A, J is positive definite, if and only if the pair fi, J is negative
definite.
P r o o f By
J H -T(A~, J ) J H r = (A, J)
the two pairs have always the same eigenvalues. By (1.6) all eigenvalues have
negative real parts. If A, J is negative definite then atl eigenvalues are real and
negative and there exists a Z with
(3.6) z T J Z = J, z T ~ 4 z = - ~J ,
Then
The second equality in (3.8) follows easily from the second one in (3.6) (note that
PJP =J). F o r the first equality in (3.8) we have
-
= p ~ - 1 / 2 Z T H J H T Z Z - 1 J Z - T Z T H J H T Z o t - UEp
which is diagonal. N o w
pFT(A, J ) A F P = - P(-ctJ, J) = ZT(--/L --J),
5 It is a little annoying that before stopping one void cycle must be run with some n2/2 scalar
products in order to check (3.3). However, this increases the whole operational count by not more
than 4-5%.
A Jacobi eigenreduction algorithm for definite matrix pairs 255
and
Z TAZ = ~ 1/2 p F T j F p a i/2 = _ aJ .
Now the algorithm runs on ,4, given by the factor H which upon con-
vergence and J-normalization yields the eigenvector matrix H ~ - 1 / 2 =
H Z a - 1/2p for the original pair. All remaining details are as in case M above. 6
The stability problems in computing H in (3.4) are obviously of a similar nature
as those with A in (1.11). However, H contains only the factors L l, L 2, L 3 and an
improvement in accuracy can be expected if these factors are available without
previously forming K, M, D [1] or are computed by using higher precision arith-
metic in the Cholesky decompositions. In this case our implicit method reminds of
a singular value decomposition and may share its advantages in accuracy, if the
matrices K, M, D have high condition numbers (see Example 4.6 below).
The operational count for obtaining H is 5(n/2) 3/6 = 5n3/48 which is about 14
times less than that of one Jacobi cycle. Since one Jacobi process takes generally
not less than 6-8 cycles the cost for computing H can be considered as negligible.
There is one interesting aspect of our implicit algorithm. The transition
from the matrix A = H J H T to d = H T j H is nothing else but one step of a
"J-symmetric" LR method for the pair A, J. This carries usually some non-
negligible diagonalization effect. Indeed, according to (1.11) we have
which is always less than TrA (note that the definiteness sign changes with the
transition from A t o / ] ! ) .
Similar remarks hold for the transition from a single matrix A = GJG r to the
auxiliary pair GTG, J in case ~r This effect is stressed by the fact that the indefinite
symmetric decomposition A = GJG T is made with total pivoting. This has some
further favorable stability consequences ([33]).
6 The implicit method is much more difficult to design for matrix pairs A, J where A and J are
both indefinite and have different inertiae. In fact, in this case it is easily shown that the
definiteness of the original pair does not carry over to the auxiliary one.
256 K. Veseli6
4. Experimental results
Ia,s [ < l O m a c h e p s ~
was satisfied for all r < s. The above inequality was also used as a threshold within
the row-cyclic strategy.
The hyperbolic parameter y was bounded according to (2.20) with tmax 0.8. =
7 In fact, we use two different versions of that algorithm: one for case ~1 and one for case ~ in
Sect. 4. There will be no confusion in using the same name for both.
s Our time experiments are generally made on matrices, generated in some random way. If the
eigenvalues are about to be separated by the Gershgorin circles Jacobi methods become faster.
This could be efficiently exploited ff approximate eigenvectors are known and are used for
a pre-transformation of the matrix.
A Jacobi eigenreduction algorithm for definite matrix pairs 257
4.1.1
I -10-101~ 1 1 1
A= 1 1 1 .
1 1 -1.0.10 s
On such matrices (in [2] they are called scaled diagonally dominant) both Jacobi
methods are good. It is interesting to note that on many matrices the accuracy ofql
or even standard Jacobi, may strongly depend on the permutaton of rows and
columns of the initial matrix. On the present example, however, no permutation
helped ql to compute one single digit of the smallest eigenvalue correctly whereas
jsmp and jsm were insensitive to such permutations.
4.1.2
I
11 11 a1 11 1
A= 1 a 1 1 "
I 1 1 b
in all cases. Also, the small eigenvalue 1 + O(1/b) was always correct withjsmp. The
accuracy ofjsm was excellent for some permutations, independently of the choice
(i), (ii), (iii) and catastrophic for other permutations. The situation with ql was
amazing: in both (i) and (ii) there was exactly one permutation yielding correct
small eigenvalue, but these permutations were different. For (iii) no permutation
helped ql to get any correct digit?
4.2
A=
[110 ]
1 0
1 0 I0-7
The well-behavedness of this matrix is partly based on its zero pattern ([38])
and it has no positive definite analogue. Here jsmp got no correct digit of the tiny
eigenvalue. This is not surprising, if one remembers that the indefinite symmetric
decomposition [5] begins by a 1 x 1 Gaussian elimination step on the pivot 1, 1
which has disastruous consequences. If [5] is told to make a 2 x 2 step instead, then
everything is accurate. It seems to be a non-trivial open problem to modify the
pivoting strategy of the indefinite symmetric decomposition to cover all such cases
(cf. also [33]).
-1 - 12 "'-
K = I 2 1
9 ~ " " --1
-1 2
If all ci are equal then both the eigenvalues and the eigenvectors of A are exactly
known (here both A -1 and the corresponding H - x are generated by the explicit
formulae for the Cholesky factors). The computing times were
cd~=4 c~=6
n ~m ~mf jsmp qr jsm ~mf jsmp qr
20 1.8 1.2 1.2 1 1.6 1 1 1
30 2 1.2 1.1 1 1.9 1.1 0.9 1
50 2.3 1.3 1.2 1 2.2 1.2 1.1 1
9 Although our experienceis limited to low dimensions,ql seemsto produce often "unexpectedly"
accurate small eigenvaluesfor some initial permutations (we see no simple rule to single out that
permutations), if the condition number does not exceed 1/macheps.Somewhere above this limit
the accuracy is worsening rapidly.
A Jacobi eigenreduction algorithm for definite matrix pairs 259
Note that here we were not able to improve the accuracy of qr results by
lowering its tolerance epsilon.
The P C M A T L A B QZ algorithm was here equally good. It yielded
-8.10108123391679.10 -7 -6.90250922120189.10 -6
which shows that the eigenvalue 0 is defective only if W r J W i s singular (and this is
"almost never" the case). Since all non-vanishing eigenvalues are non-defective (i)
follows. Now, if the pair W W x, J were positive definite, then according to Th. A.1
the form xX,lx would be definite on the nullspace of W W r (which is "almost
always" n - r-dimensional). By the Sylvester inertia theorem this is impossible, if
n - r > m and n - r > n - m. Thus, (ii) holds.
260 K. Veseli6
We have tried a total of 3630 matrices for n = 10, 20, 30, 50 and various choices
of r, m, r/producing matrices from the boundary of the definiteness region or its
small neighbourhood (r/= 20i. macheps, i = 0, 1, 2, 3). The tested matrices were
distributed over the dimension n as follows
dimension n 10 20 30 50
number of tested matrices 1680 670 360 220
We have measured two condition numbers of the current eigenvector matrix
Z = (zl . . . . . z~):
IlZll~ and llZll~ = max[lz/l[ 2
Here obviously
1 < [tZllz < IlZ[l~, n < llZ[l~
Possible measures for the stability of the eigenvector computation are the
quantities
maximal l[Z II~ maximal llZ I[
finalllZ[l~ " ~= finaIllZIl~
The following table shows the maximum values of the measured (~, (l:
n 10 20 30 50
(E 2 4.4 2.8 1.4
~t 4.25 9 10.1 6.2
(the final conditions themselves never exceeded some hundreds i.e. the generated
matrices had well conditioned eigenvectors). On all other tested types of matrices
(e.g. those with r = n) these quotients were always about one. On matrices
A = GrG obtained by (3.1) with total pivoting the danger of high condition is
essentially smaller and well-controlled (s. [33]).
We also have tested a number of matrices with high eigenvector condition (they
were obtained by taking r = n thus producing a positive definite A and then fitting
a negative ~/so as to bring the pair A, J to the boundary of the definiteness region).
In all such cases both values (E, (t were 1. The results also fully confirmed the
perturbation estimate (A. 11) from Th. A.4. Moreover, we have also compared the
relative accuracy of the computed QR and Jacobi eigensolutions in single precision
(macheps,~lO-S). The comparison was made against extended precision
(macheps ~ 10- 20) results of both QR and Jacobi. Here Jacobi was hardly essen-
tially worse and often better than QR - - irrespectively of the condition of the
eigenvector matrix.
n j s m f jsmp qr
20 1.1 1 1
30 1.2 1.1 1
50 1.4 1.2 1
4.6
l+e 0.5 0
Mo=
Ei ~176 ~l
0.5
0
2+e
0
0
e
'
qr ~mf jsmp o
-1.99002011775738-10 - I - -1.99002011664866-10 - 1 --1.99002011664867.10 - I
-3.75778707876245.10-1 -- 3.75778706961464.10 - 1 --3.75778706961460.10 -1
-5.00059269396540.10 -1 - -5.00059269396839.10 - 1 --4.99999999999999.10 -1
- 1.00000000000000.10 -1 - -9.99999999999975.10 - 1 --9.99999999999975.10 -1
-1.62422129292789 -- 1.62422129303876 -- 1.62422129303876
-1.80099798742062 - -1.80099798833540 --1.80099798833542
--1.99994073060351 - 1.99994073060316 - - 2 . ~ 2
-2.(X)OOOXXXX)O090.1013 -- 2.00000000000090- I013 --2.000000(0)O30090.1013
M , D, K are as follows
3 2 2
M=I, K= 2 3 2 ' D= 1 2 1,
2 2 3 1 1
w i t h e 1 = 1 . 1 1 . 1 0 - 12. T h e e i g e n v a l u e s are e x a c t l y k n o w n a n d up to an e 2 r e l a t i v e
e r r o r t h e y are e q u a l to - 5/e ~, - 1/e ~ (triple), - e 1 (triple), - 9 e l / 5 . H e r e we
h a v e c o m p a r e d several P C M A T L A B r o u t i n e s w i t h jsmf. W e h a v e tried the m a t r i x
A~ = A -1 f r o m (1.11) u s i n g P C M A T L A B a n d its f u n c t i o n chol. W e t h e n c o m -
p a r e d v a r i o u s P C M A T L A B Q R / Q Z possibilities w i t h o u r jsmf a l g o r i t h m . T h e
results b e l o w a r e labeled as follows
H e r e T, S are t h e m a t r i c e s f r o m (1.9). T h e o b t a i n e d e i g e n v a l u e s w e r e
Here only jsmf was able to reproduce both the small and the large eigenvalues
correctly. Note also that neither A nor A - 1 is diagonally d o m i n a n t in the sense of
[2] thus indicating further classes of matrices on which Jacobi methods are
accurate.
In this section we present some theoretical results mainly concerning the eigen-
vector condition of definite pairs. The results are mostly new. Although useful in
c o n n e c t i o n with our algorithm, they may have i n d e p e n d e n t interest. A particularly
strong result is T h e o r e m A.4 which describes the behaviour of the condition of the
eigenvectors of a pair A, J near the b o u n d a r y of the definiteness region. We prove
that the c o n d i t i o n n u m b e r is O(1/x/d), d the distance to the nearest indefinite pair.
Thus, although possible, the b a d condition is highly i m p r o b a b l e and, which is
much more i m p o r t a n t , it can be sensibly improved by only a slight change of the
matrix.
The first two assertions of the following theorem are elementary a n d their proof
is included for the sake of completeness. The last assertion is i m p o r t a n t a n d seems
to be new.
Theorem AI. Let the pair A, B be, say, positive definite and let B be non-singular.
Then
(i) There exists a n o n - s i n g u l a r Z such that
where A + . diag(2~-,
. . . . 2,,+ ), A _ = diag(21- . . . .2~--m)
. . and
(A.2) j = ( lm O)
0 -- In- m
where
Proof. Let # be any definitizing shift. Then as it is well known there exists a non-
singular Z 1 such that
ZT(A -- l , B ) Z l , fl = Z T B Z l
are diagonal where the first m diagonal elements of fl are positive and the remaining
ones are negative. Taking Z = Z 1 ]ill - 1/2 we obtain Z T B Z = J and ZT(A -- # B ) Z
diagonal. Then
ZTAZ = 0 -A _
is diagonal as well, which proves (i). (ii) follows from (i) and the positive definiteness
of
ZT(A_#B)z=(A+ -# 0 ) .
0 -A-+I~
N o w we prove (iii). Let Z o and Z perform (A.1) i.e.
ZTAZo =
0)
-A o , Z T B Z o = J, A~:diagonal.
,A3, _Ao):OT( o )o
By the J-orthogonality of D we have
o(T0
where by the J-orthogonality of D the matrices U 1, U 2 are orthogonal. Thus
(U~ 0)
Z = Zo 0 U2
Corollary A2. All matrices Z performing the transformation (A.1) for a definite pair
A, B have the same (Euclidean, spectral) condition number.
Theorem A3. Let the condition of Theorem A.1 be fullfilled with B = J. Then the
Euclidean condition of Z is bounded by
(A.5) c o n d E Z = flZllEIIZ -x II~ -< n + 2p(A) ,
with
~, n-m
(A.6) tro = T r A + - T r A _ = 2+ - ~, 27
i=1 i=l
Proof. By the J-orthogonality of Z the left hand side of (A.5) reads simply T r Z T Z .
A straightforward computation shows that the polar decomposition of any
J - o r t h o g o n a l matrix reads
x/I + X X T X )
(A.8) H(X) = XT x/I + x T x ,
+ 0 )Z -1
(A.9)
-Z_
where
A+ = diag(2 + . . . . . 2+),
(A.10)
A _ = diag(2i- . . . . . 2~,_,).
Therefore
T r A = T r [ x / I + X X T A +x/I + X X T -- X A _ X T + X T A +X
The formula (A.5) is, of course, only an upper estimate for the condition
number of Z and there are matrices for which this estimate is pessimistic. In such
cases it is likely that a non definite pair A 1, J could be found in a small neigbour-
hood of A, J.
In the following theorem we derive an asymptotic estimate of the condition of
the diagonalizing J-orthogonal matrix as A approaches the boundary of the
definiteness region. Without loss of generality we can assume that the spectral
point where the plus and the minus eigenvalues come together is zero.
Theorem A4. Let A and V be positive semidefinite such that the definiteness interval
of A, J is void (this implies that A is singular). Let, in addition, A + V be positive
definite or, equivalently,
Ax=O ~ xTVx>O.
Then A + ~ V is positive definite for any e > 0 and the spectral condition of any
J-orthogonal matrix Z diagonalizing the pair A + e V, J is estimated by
A/ll 0 AP12
(A.12) FT(A + eV)F = 0 A'oo 0
A ?2 0 A ~22
and the pair A~ Jo has only one multiple eigenvalue zero. Here
J = diag(I, Jo, - I ) ,
is partitioned as in (A.12). The existence of such an F is guaranteed by the analytic
perturbation theory (see Kato [24], Ch. II w where F is called the transforma-
tion function). For e > 0 Abo is positive definite and the pair Abo, Jo is diagonal-
ized by a matrix Zo, obtained as follows. We take L with A~,o = L L T and find an
orthogonal U such that
UTL-1JoL-T U = a-lJo ,
we obtain
'7 T A ! "7
Z~JoZo = Jo, Z,o,aooZ, o = ~ .
We have IIL-Tll = I I Z ~ 1 I11]2 The smallest eigenvalue of Abo is analytic at e = 0
and reads
where
yTVy
v= inf
Ay=0 YT(F ~ 1 7 6 -ly > 0 .
On the other hand, let/~ be any of the diagonal elements of ~. Then/~ > 0 and
(note that A~ - #Jo is invertible for every #:~0). Since d o A ~ is nilpotent and
A ~ is positive semidefinite it is readily seen that already the square of J o A ~ or,
equivalently, the matrix A oooJoA ooo vanishes. Thus,
Jo JoA~
(A.14) (A~ - # S o ) - 1 = _ __ +
# #:
F r o m this we obtain
/~ < IIAtl + Ilf~
and thus,
t)~1/211 ~ ~IIAII + IIF~ ~A-HIIAII.
Putting
Z = F diag(I, Zo, I)
we obtain
z T J Z = J, ZT A ' Z = d i a g .
Now
Theorem AS. The system described by (1.6) is overdamped if and only if the pair A, J
from (1.11) is positive definite.
Proof. It is well k n o w n (see e.g. [10]) that (1.7) holds, if and only if there exists
a real # such that
--(M/,t 2 + D,u + K)
[, -- btI 0 -~2M
0
+/~D + K)
]['o ;']
Thus, the positive definiteness of the pair S, T and therefore the positive definite-
ness of A, J is equivalent to the overdampedness condition (1.8). Q.E.D.
Acknowledgements. The first version of this article is more than six years old and during this time
it has incited several follow-on papers I-8,9, 33, 38] which, in turn, influenced our present version.
In addition, Stapni~ar, in preparing his thesis 1-331made a first implementation of our eigenreduc-
tion in case ~r above. This implementation was used in our experiments. The author thanks the
authors of these papers for early communicating their results and experiences. He also thanks I.
Slapni~ar, Hagen and A. Suhadolc, Ljubljana, for many valuable discussions.
References
1. Argyris, J.H., Bronlund, O.E. (1975): The natural factor formulation of stiffness for the matrix
deplacement method. Comput. Meth. Appl. Mech. Engrg. 40, 97-119
2. Barlow, J., Demmel, J. (1990): Computing accurate eigensystems of scaled diagonally
dominant matrices. SIAM J. Numr. Anal. 27, 762-791
3. Brent, R.P., Luk, F.T. (1985): The solution of singular-value and symmetric eigenvalue
problems on multiprocessor arrays SIAM J. Sci. Statist. Comput. 6, 69-84
4. Brent, R.P., Luk, F.T., van Loan, C. (1985): Computation of the singular value decomposition
using mesh-connected processors. J. VLSI Computer Systems 1, 242-270
5. Bunch, J.R., Parlett, B.N. (1971): Direct methods for solving symmetric indefinite systems of
linear equations. SIAM J. Numer. Anal. 8, 639-655
6. Chartres, B.A. (1962): Adaptation of the Jacobi Method for a Computer with Magnetic Tape
Backing Store. Comput. J. 5, 51-60
7. Demmel, J., Veselir, K. (1992): Jacobi's method is more accorate than QR, Courant Institute
preprint 1989. SIAM J. Matr. Anal. Appl. (to appear)
8. Donner, K. (1990): Uber ein Jacobi-Verfahren ftir definite hermitesche Matrix-Paare. Fern-
universit/it Hagen, Diplomarbeit
9. Drma~, Z., Hari, V. (1993): On quadratic convergence bounds for the J-symmetric Jacobi
method. Numer. Math. (to appear)
10. Duffin, R.J. (1955): A minimax theory for overdamped networks. J. Rat. Mech. Anal. 4,
221-233.
11. Durand, E. (1972): Solutions numeriques des equations algrbriques. Masson, Paris
12. Eberlein, P.J., Boothroyd, J. (1968): Solution to the eigenprobtem by a norm reducing Jacobi
type method. Numer. Math. 11, 1-12 (also [41-] p. 327)
13. Falk, S., Langemeyer, P. (1960): Das Jacobische Rotationsverfahren fiir reellsymmetrische
Matrizenpaare I, II. Elektronische Datenverarbeitung, pp. 30-43
A Jacobi eigenreduction algorithm for definite matrix pairs 269
14. Falk, S. (1984): Ober halbimplizite und implizite Algorithmen zur endlichen und iterativen
Transformationen von Matizenpaaren. ZAMM 64, 437439
15. Golub, G.H., van Loan, Ch.F. (1989): Matrix computations. Hopkins University Press,
Baltimore
16. Gose, G. (1979): Das Jacobl-Verfahren ffir A x = 2 B x . ZAMM 59, 93-101
17. Hari, V. (1982): On the global convergence of the Eberlein method for real matrices. Numer.
Math. 39, 361-369
18. Hari, V. (1984): On cyclic Jacobi methods for the positive definite generalized eigenvalue
problem. Fernuniversitht Hagen, Dissertation
19. Hari, V.: On the global convergence of cyclic Jacobi methods for the generalized eigen-
problem. Department of Mathematics, University of Zagreb, preprint
20. Haft, V. (1991): On pairs of almost diagonal matrices. Linear Algebra Appl. 148, 193-223.
21. Hari, V., Veseli6, K. (1987): On Jacobi methods for singular value decompositions. SIAM J.
ScL Stat. Comp. 8, 741-754
22. Hestenes, M.R. (1958): Inversion of matrices by biorthogonalization and related results.
J. STAM 6, 51-90
23. Kaiser, H.F. (1972): The JK method: a procedure for finding the eigenvalues and eigenvectors
of a real symmetric matrix. Compout. J. 15, 271-273
24. Kato, T. (1966): Perturbation theory for linear operators, Springer, Berlin Heidelberg
New York
25. Luk, F.T., Park, H. (1987): A proof of convergence for two parallel Jacobi SVD algorithms.
Proc. SPIE 826. Advanced algorithms and signal processing II, San Diego, CA
26. Moler, C.B., Stewart, G.W. (1973): An algorithm for generalized matrix eigenvalue problems.
SIAM J. Numer. Anal. 10, 241-256
27. Nash, J.C. (1974): A one-sided transformation method for the singular value decomposition
and algebraic eigenvalue problem. Comput. J. 18, 74-76
28. Rath, W. (1982): Fast Givens rotations for orthogonal similarity transformations. Numer.
Math. 40, 47-56
29. Sameh, A.H. (1971): On Jacobi and Jacobi-like algorithms for a parallel computer. Math.
Comput. 25, 579-590
30. Scott, D.S., Ward, R.C. (1982): Solving symmetric-definite quadratic 2-matrix problems
without factorization. SIAM J. Sci. Star. Comput. 3, 58-110
31. Shroff, G., Schreiber, R. (1988): On the convergence of the cyclic Jacobi method for parallel
block orderings. Rensselaer Polyt. Inst. Dept. Comput. Sci. Report No. 88-11
32. Stapni6ar, I., Hari, V. (1991): On the quadratic convergence of the Falk-Langemeyer method.
SIAM J. Math. Anal. 12, 84-114
33. Slapni6ar, I. (1992): Accurate symmetric eigenreduction by a Jacobi method, Fernuniversit~it
Hagen, Ph.D. Thesis
34. Stewart, G.W. (1979): Perturbation bounds for the definite generalized eigenvalue problems.
Linear Alg. Appl. 23, 69-85
35. Veseli6, K. (1976): Some convergent Jacobi-like procedures for diagonalizing J-symmetric
matrices. Numer. Math. 27, 67-75
36. Veseli6, K. (1983): Global Jacobi method for a symmetric indefinite problem S x = 2 T x
Comput. Meth. Appl. Mech. Engrg. 38, 273-290
37. Veseli6, K., Haft, V. (1989): A note on a one sided Jacobi algorithm. Numer. Math. 56,
627-633
38. Veseli6, K., Slapni6ar, I. (1992): Floating-point perturbations of Hermitian matrices. Fern-
universitat Hagen, preprint
39. Wang, X., (t991): A few propositions on upper bounds for max 0 ~ m~ M x ( A , , , ) / x ( A o ) . Fern-
universitht Hagen, preprint
40. Wilkinson, J.H. (1968): Almost diagonal matrices with multiple or close eigenvalues Linear
Algebra Appl. 1, 1-12
41. Wilkinson, J.H., Reinsch, C. (1971): Handbook of Automatic Computation II Linear Algebra.
Springer, Berlin Heidelberg New York
42. Zimmermann, K. (1965): On the convergence ofa Jacobi process for ordinary and generalized
eigenvalue problems. ETH Zfirich Dissertation No. 4305