Documentos de Académico
Documentos de Profesional
Documentos de Cultura
In this chapter we will study inequalities that are used for localising the
spectrum of a Hermitian operator. Such results are motivated by several
interrelated considerations. It is not always easy to calculate the eigen-
values of an operator. However, in many scientific problems it is enough
to know that the eigenvalues lie in some specified intervals. Such infor-
mation is provided by the inequalities derived here. While the functional
dependence of the eigenvalues on an operator is quite complicated, several
interesting relationships between the eigenvalues of two operators A, Band
those of their sum A + B are known. These relations are consequences of
variational principles. When the operator B is small in comparison to A,
then A + B is considered as a perturbation of A or an approximation to A.
The inequalities of this chapter then lead to perturbation bounds or error
bounds.
Many of the results of this chapter lead to generalisations, or analogues,
or open problems in other settings discussed in later chapters.
ponents >'i(A), ... ,>';(A), where >';(A) are arranged in decreasing order;
i.e., >';(A) = aj. Similarly, >.r(A) will denote the vector with components
>.} (A) where >.} (A) = an-j+I! 1 ~ j ~ n.
Theorem 111.1.1 (Poincare's Inequality) Let A be a Hermitian operator
on fi and let M be any k-dimensional subspace offi. Then there exist unit
vectors x,y in M such that (x,Ax) ~ >.i(A) and (y,Ay) ~ >.r(A).
Proof. Let N be the subspace spanned by the eigenvectors Uk, ... ,Un of
A corresponding to the eigenvalues >'hA) , ... , >.;(A). The~
dim M +dimN = n+ 1,
and hence the intersection of M and N is nontrivial. Pick up a unit vector
n n
X in M nN. Then we can write x = L~jUj, where LI~jI2 = 1. Hence,
j=k j=k
n n
(x, Ax} = LI~jI2>';(A) ~ LI~jI2>.i(A) = >'i(A).
j=k j=k
This proves the first statement. The second can be obtained by applying
this to the operator -A instead of A. Equally well, one can repeat the
argument, applying it to the given k-dimensional space M and the (n -
k + I)-dimensional space spanned by Ul, U2,· .. , Un-k+l. •
This is not surprising. More surprising, perhaps even shocking, is the fact
that we could have Ai(A) = min{(x, Ax) : x E M, Ilxll = I}, even for a
k-dimensional subspace that is not spanned by eigenvectors of A. Find an
example where this happens. (There is a simple example.)
A=( ~ :).
We say that B is the compression of A to the subspace M.
Corollary 111.1.5 (Cauchy's Interlacing Theorem) Let A be a Hermitian
operator on 1-£, and let B be its compression to an (n - k)-dimensional
subspace N. Then for j = 1,2, ... ,n - k
(Ill. 1)
Proof. For any j, let M be the span of the eigenvectors VI, ... , v j of B
corresponding to its eigenvalues Ai(B), ... , A;(B). Then (x, Bx) = (x, Ax)
for all x E M. Hence,
and
d (B)
-Aj \ i (B)
= Aj - = d
A(n-k)-j+1 (- B) for all 1 <_ J' :s: n - k.
60 III. Variational Principles for Eigenvalues
Choose i = j+k. Then the first inequality yields-A; (B) ::; -A;+k(B), which
is the second inequality in (IIl.1). •
Proof. Let AUj = O'.jUj; then {Uj} constitute the standard orthonor-
mal basis in en. There is a one-to-one correspondence between (n - 1)-
dimensional orthogonal projection operators and unit vectors given by
P = 1- zz*. Each unit vector, in turn, is completely characterised by
its coordinates (j with respect to the basis Uj. We have z = .L (jUj =
.L(u~z)Uj,.L l(jl2 = 1. We will find conditions on the numbers (j so that,
for the corresponding orthoprojector P = I - zz* , the compression of A to
the range of P has eigenvalues f3i'
Since PAP is a Hermitian operator of rank n - 1, we must have
n-1
II (A - f3i) = tr I\n-1 [P(>.I - A)P].
i=1
=L II (>. -
n
I\n-1(>.I - A) O'.k) I\n-1 E j .
j=1 k¥-j
The interlacing property ensures that all'j are nonnegative. Now choose
(j to be any complex numbers with !(j!2 = 'j. Then the equation (IlL3) is
satisfied for the values A = aj, 1 ::; j ::; n, and hence it is satisfied for all A.
Comparing the leading coefficients of the two sides of (IlL3), we see that
L!(jj2 = 1. This completes the proof.
j •
III.2 Weyl's Inequalities
Several relations between eigenvalues of Hermitian matrices A, B, and A + B
can be obtained using the ideas of the previous section. Most of these results
were first proved by H. Weyl.
Theorem III.2.1 Let A, B be n x n Hermitian matrices. Then,
A;(A + B) ::; (x, (A + B)x) = (x, Ax) + (x, Bx) ::; AhA) + A;_i+l(B).
111.2 Weyl's Inequalities 63
Note that DD* = DIDi + D2D'2. Now the nonzero eigenvalues of the
matrix C = D* D are the same as those of DD*. The same is true for the
matrices A = Di Dl and Dl Di, and also for the matrices B = D'2 D2 and
D2D'2. Hence, using Weyl's inequality (IIl.8) we get (III.9) in this special
case.
I[ In I- 0, subtract InI from C. Then all eigenvalues of A, B, and Care
translated by -In' By the special case considered above we have
WI, W2, ... , Wk-I and hence be an element of U. So, span{ WI, V2,···, vd is
a k-dimensional space that must, therefore, be U +span{v}. Now WI E WI
and Vj E W j , j = 2, ... ,k. Again all requirements are met. •
Hence,
dim 8 j :::: ik - i j + 1 :::: k - j + l.
Note that 8 1 :J 8 2 :J ... :J 8 k are subspaces of Vk and Wj E 8 j for j =
1,2, ... ,k-l. Hence, by Lemma III.3.1 there exists a vector u in 8 1 -U such
.
that the space U +span{ u} has a basis UI, ... , Uk, where Uj E 8 j c Wj,j =
1,2, ... , k. But U + span{u} is also the linear span of VI, ... , Vk-I and u.
Put Vk = u. Then Vj E Vi, j = 1,2, ... ,k, and they span the same space as
~~.
(ii) Let Wj = span{ Uj, . .. ,un}, 1 :::; j :::; n. Choose orthonormal vectors
Xi} from the spaces W ij , j = 1, ... ,k. Let W be the span of these vectors
III.3 Wielandt's Minimax Principle 67
This proves (i). The statement (ii) has exactly the same proof.
•
Theorem 111.3.5 (Wielandt's Minimax Principle) Let A be a Hermitian
operator on an n-dimensional space 1t. Then for any indices 1 ::; i1 < ... <
ik ::; n we have
k k
2:);j(A) max
.M,C··C.Mk
min
XjEMj
2:(Xj, AXj)
j=l dim Mj=ij xj orthonormal j=l
k
min max 2:(Xj,AXj).
N,~···~Nk XjEN'j
dim Nj=n~ij+l xj orthonormal j=l
Proof. We will prove the first statement; the second has a similar proof.
Let Vij = span{ U1, ... ,Ui j }, where, as before, the Uj are eigenvectors of A
corresponding to AJ(A). For any unit vector X in Vij , (x,Ax) ;:::: A;j(A). So,
if Xj E Vij are orthonormal vectors, then
k k
Hence , the desired result will be achieved if we prove that given any sub-
spaces M1 C ... C Mk with dim M j = i j we can find orthonormal vectors
Xj E M j such that
k k
Let Nj = Wij = span{vij, ... ,vn},j = 1,2, ... ,k. These spaces were
considered in Proposition III.3.4(ii). We have Nl ::J N2 ·•• ::J Nk and
dimNj = n - ij + 1. Hence, by Theorem III.3.2 and Exercise III.3.3 there
exist orthonormal vectors Xj E M j and orthonormal vectors Yj E Nj such
that
span{xl, ... , xd = span{Yl, ... , Yk} = W, say.
By Proposition III.3.4 (ii), A~(Aw):$ A;)A) for j = 1,2, ... ,k. Hence,
k k
~)xj,Axj) ~)xj,Awxj} = tr Aw
j=l j=l
k k
= LA~ (Aw) :$ LAt (A).
j=l j=l
This is what we wanted to prove. •
Exercise 111.3.6 Note that
k k
LAUA) = L(Uij,Auij}.
j=l j=l
We have seen that the maximum in the first assertion of Theorem III.3.5
is attained when M j = Vij = span{ul, ... ' UiJ,j = 1, ... , k, and with this
choice the minimum is attained for Xj = Uij' j = 1, ... , k. Are there other
choices of subspaces and vectors for which these extrema are attained? (See
Exercise III.l.3.)
Exercise 111.3.7 Let [a, b] be an interval containing all eigenvalues of A
and let <1>( tl, ... , tk) be any real valued function on [a, b] x ... x [a, b] that is
monotone in each variable and permutation-invariant. Show that for each
choice of indices 1 :S il < ... < ik :$ n,
max
MIC···CMk w=spa~!~ ... 'Xk} <I> (Ai(A W ), ... , AhAw)) ,
dim Mj=i j xjEMj.xj orthonormal
k k k
L>.t(A+B):::; L>';j(A) + L>.;(B). (III.10)
j=l j=l j=l
k k
k k
for any choice of orthonormal vectors Xl, ... , xk. The above two relations
imply that
L>';j(A+B) :::;
j=l
Now, using Theorem III.3.5 once again, it can be concluded that the first
J.
k
term on the right-hand side of the above inequality is dominated by L>'; (A).
j=l
(III. 11 )
Exercise 111.4.3 (Lidskii's Theorem) The vector>.l(A+B) is in the con-
vex hull of the vectors >.1 (A) + p>.l(B), where P varies over all permuta-
tion matrices. (This statement and those of Theorem III.4.1 and Corollary
111.4.2 are, in fact, equivalent to each other.)
Lidskii's Theorem can be proved without calling upon the more intricate
Wielandt's principle. We will see several other proofs in this book, each
highlighting a different viewpoint. The second proof given below is in the
spirit of other results of this chapter.
70 III. Variational Principles for Eigenvalues
Case 1. ik < n. Let M = span{ WI, ... ,Wn -l} and let AM be the compres-
sion of A to the space M. Then, by the induction hypothesis
k k k
LAJj(A M +BM ) ~ LAr,(A M ) + LA;(BM ).
j=l j=l j=l
The inequality (111.10) follows from this by using the interlacing principle
(111.2) and Exercise IIL1.7.
Case 2.1 < i 1· Let M = span{u2, ... ,un}. By the induction hypothesis
k k k
LAr,-l(A M + B M ) ~ LAr,-l(A M ) + LA;(BM ).
j=l j=l j=l
Once again, the inequality (111.10) follows from this by using the interlacing
principle and Exercise 111.1. 7.
Case 3. il = 1. Given the indices 1 = il < i2 < ... < ik ~ n, pick up the
indices 1 ~ £1 < £2 < ... < £n-k < n such that the set {i j : 1 ~ j ~ k}
is the complement of the set {n - £j + 1 : 1 ~ j ~ n - k} in the set
{I, 2, ... ,n}. These new indices now come under Case 1. Use (111.10) for
this set of indices, but for matrices -A and - B in place of A, B. Then note
that A;( -A) = -A~_j+l (A) for alII ~ j ~ n. This gives
n-k n-k n-k
L - A~_tj+1 (A + B) ~ L- - A~_tj+1 (A) + L - A~_j+l (B).
j=l j=l j=l
Now add tr(A + B) to both sides of the above inequality to get
k k k
LAr,(A + B) ~ LAr,(A) + LA;(B).
j=l j=l j=l
This proves the theorem.
•
As in Section 111.2, it is useful to interpret the above results as pertur-
bation theorems. The following statement for Hermitian matrices A, B can
be derived from (III.11) by changing variables:
(111.12)
IlIA Lidskii's Theorems 71
for all indices 1 :S i1 < ... < ik :S n. Note that we are allowing the
possibility of zero coordinates in this notation.
Theorem 111.4.5 (Gel'Jand-Naimark) Let A, B be any two operators on
H. Then the singular values of A, Band AB satisfy the majorisation
log s(AB) - log s(B) -< log s(A). (III. 17)
Proof. We will use the result of Exercise III.3.7. Fix any index k,1 :S
k :S Choose any k orthonormal vectors Xl, ... ,Xk, and let W be their
n.
linear span. Let q,(h, ... , tk) = ht2··· tk· Express AB in its polar form
AB = UP. Then, denoting by Tw the compression of an operator T to the
subspace W, we have
q, (Ai(Pw), ... , A%(PW )) 1 det Pwl 2
1det( (Xi, PWXj)) 12
Idet((Xi,PXj))1 2
The second of these determinants is equal to det(B* B)w; the first is equal
k
to det(AA*)uw and by Corollary IIL1.5 is dominated by II s;(A). Hence,
j=l
we have
k
<I> (A~(PW)' ... ' A%(Pw )) ::; det(B* B)w II s;(A)
j=l
k
= <I> (A1(IBI~),· .. , Ak(IBI~)) II s;(A).
j=l
Now, using Exercise IIL3.7, we can conclude that
k k k
(II A;j (p))2 ::; IIAt (IBI2) IIs;(A),
j=l j=l j=l
i.e.,
k k k
II Sij (AB) ::; IISij (B) II sj(A), (III.18)
j=l j=l j=l
for alII::; i1 < ... < ik ::; n. This, by definition, is what (III.17) says. •
log s.l.(A) + log si(B) -< log s(AB) -< log s.l.(A) + log s.l.(B) (IIL20)
for any two matrices A, B. In conformity with our notation this is a sym-
bolic representation of the inequalities
k k k k k
Let Xl, ... ,xn be an orthonormal Schur-basis for A such that Aj(A)
= (Xj, AXj). Then Aj(A) = (Xj, A*xj). Let W = span{x1,.·., xd. Then
k
L(Xj, (Re A)xj) = tr (ReA)w
j=l
k k
IA + BI = W*(A + B) = Re W*(A + B) = Re W* A + Re W* B.
Now use Exercise 1II.5.5(ii). •