Peter Denton, Stephen Parke, Xining Zhang, and I have just uploaded to the arXiv the short unpublished note “Eigenvectors from eigenvalues“. This note gives two proofs of a general eigenvector identity observed recently by Denton, Parke and Zhang in the course of some quantum mechanical calculations. The identity is as follows:
Theorem 1 Let
be an
Hermitian matrix, with eigenvalues
. Let
be a unit eigenvector corresponding to the eigenvalue
, and let
be the
component of
. Then
where
is the
Hermitian matrix formed by deleting the
row and column from
.
For instance, if we have
for some real number ,
-dimensional vector
, and
Hermitian matrix
, then we have
assuming that the denominator is non-zero.
Once one is aware of the identity, it is not so difficult to prove it; we give two proofs, each about half a page long, one of which is based on a variant of the Cauchy-Binet formula, and the other based on properties of the adjugate matrix. But perhaps it is surprising that such a formula exists at all; one does not normally expect to learn much information about eigenvectors purely from knowledge of eigenvalues. In the random matrix theory literature, for instance in this paper of Erdos, Schlein, and Yau, or this later paper of Van Vu and myself, a related identity has been used, namely
but it is not immediately obvious that one can derive the former identity from the latter. (I do so below the fold; we ended up not putting this proof in the note as it was longer than the two other proofs we found. I also give two other proofs below the fold, one from a more geometric perspective and one proceeding via Cramer’s rule.) It was certainly something of a surprise to me that there is no explicit appearance of the components of
in the formula (1) (though they do indirectly appear through their effect on the eigenvalues
; for instance from taking traces one sees that
).
One can get some feeling of the identity (1) by considering some special cases. Suppose for instance that is a diagonal matrix with all distinct entries. The upper left entry
of
is one of the eigenvalues of
. If it is equal to
, then the eigenvalues of
are the other
eigenvalues of
, and now the left and right-hand sides of (1) are equal to
. At the other extreme, if
is equal to a different eigenvalue of
, then
now appears as an eigenvalue of
, and both sides of (1) now vanish. More generally, if we order the eigenvalues
and
, then the Cauchy interlacing inequalities tell us that
for , and
for , so that the right-hand side of (1) lies between
and
, which is of course consistent with (1) as
is a unit vector. Thus the identity relates the coefficient sizes of an eigenvector with the extent to which the Cauchy interlacing inequalities are sharp.
— 1. Relating the two identities —
We now show how (1) can be deduced from (2). By a limiting argument, it suffices to prove (1) in the case when is not an eigenvalue of
. Without loss of generality we may take
. By subtracting the matrix
from
(and
from
, thus shifting all the eigenvalues down by
, we may also assume without loss of generality that
. So now we wish to show that
The right-hand side is just . If one differentiates the characteristic polynomial
at , one sees that
Finally, (2) can be rewritten as
so our task is now to show that
By Schur complement, we have
Since is an eigenvalue of
, but not of
(by hypothesis), the factor
vanishes when
. If we then differentiate (4) in
and set
we obtain (3) as desired.
— 2. A geometric proof —
Here is a more geometric way to think about the identity. One can view as a linear operator on
(mapping
to
for any vector
); it then also acts on all the exterior powers
by mapping
to
for all vectors
. In particular, if one evaluates
on the basis
of
induced by the orthogonal eigenbasis
, we see that the action of
on
is rank one, with
for any , where
is the inner product on
induced by the standard inner product on
. If we now apply this to the
-form
, we have
, while
is equal to
plus some terms orthogonal to
. Since
, Theorem 1 follows.
— 3. A proof using Cramer’s rule —
By a limiting argument we can assume that all the eigenvalues of are simple. From the spectral theorem we can compute the resolvent
for
as
Extracting the component of both sides and using Cramer’s rule, we conclude that
or in terms of eigenvalues
Both sides are rational functions with a simple pole at the eigenvalues . Extracting the residue at
we conclude that
and Theorem 1 follows. (Note that this approach also gives a formula for for
, although the formula becomes messier when
because the relevant minor of
is no longer a scalar multiple of the identity
.)