Lecture 18

[[lecture-data]]

2024-10-09

Readings

4. Chapter 4

Interlacing II / Inclusion Principle

Suppose AMn is hermitian and B is an r×r principal submatrix. Then for all k,

λk(A)λk(B)λk+nr(A)
Proof (via Courant-Fisher)

Say B comes from A, deleting rows and columns i1,i2,,inr.

λk(A)=maxy1,y2,,yk1CnminxCn0,xyiixAxxx=maxy1,,yk1minxyii,xei1,,einrxAxxx()=maxv1,v2,,vk1CrminzCr0,zv1,,vk1zBzzz=λk(B) by Courant-Fischer
  • Note that the eik are the standard basis vectors with the 1 in the index of each ik
  • The zs are the xs without the ik components (since they are orthogonal to those standard basis vectors)
  • We can then "perform surgery" on the ys also to get rid of those coordinates to get v1,,vk1Cr

(see inclusion principle)

Corollary

Suppose AMn hermitian and A^ is a n1×n1 principal submatrix. Then

λ1(A)λ1(A^)λ2(A)λ2(A^)λ3(A)λk(A^)λk(A)

(see the eigenvalues of a hermitian matrix and a principal submatrix one smaller alternate in magnitude)

Corollary

For any AMn hermitian, i we have λ1(A)aiiλn(A).

This follows immediately from the inclusion principle, since each diagonal entry is a 1×1 principal submatrix

(see the diagonal entries of a hermitian matrix are bounded by the eigenvalues)

Majorization

Let x,yRn. Say we can order the components of x so that x1xn and the same for the components of y so we have yk1ykn. We say x majorizes y precisely when

  • For all m,i=1mxii=1myki and
  • equality holds when m=n

(see majorization)

Theorem

Let AMn be hermitian. Then the vector of diagonal elements of A, call it diagA majorizes the vector of ordered eigenvalues of A, call it Λ(A)

Proof via induction on A

Any case where n=1 is trivially true. Assume the theorem holds for all matrices of size n up to some fixed k1. Consider the case when n=k.

Let B be a submatrix of A obtained by deleting one row and corresponding column k where the diagonals of A are ordered a11a22akk.

For all n=1,,k1, we have i=1kλi(A)i=1kλi(B) by interlacing 2. Then by the induction hypothesis, we have that i=1nλi(B)i=1naii. But for the case when n=k, we have that

i=1nλi(A)=Tr(A)=i=1na

Thus the theorem holds

(see diagonal elements of a hermitian matrix majorize its eigenvalues)

Corollary

Let AMn be hermitian and r:1rnZ. Then

k=1rλk(A)=minUMn,r orthonormal colsTrUAU

and also

k=0r1λnk(A)=maxUMn,r orthonormal colsTrUAU

And claim that this implies the previous result ( just take take U as the identity )

Proof / Intuition

Given any UMn with orthonormal columns, extend Gram-Schmidt to get V=[U|]Mn unitary. Then when we do the multiplication

[VAV]r×r=UAU

So by interlacing 2, we get that

λk(A)=λk(VAV)λk(UAU)

So sum over k=1,2,,r to get

k=1rλk(A)Tr(UAU)

So we have the desired lower bound, and need to show we can achieve equality.

Let W be the orthonormalized eigenvectors associated with λ1(A),,λr(A).

TrUAU=TrWAW=TrDr

(the sum of the first r eigenvalues)

(see the sum of the first least eigenvalues is the minimum of the trace of orthonormal multiplications)