In this appendix we use the Jordan normal form theorem to study the asymptotic dynamics of transition matrices such as those of Markov chains introduced in Section ??.

The basic result is the following theorem.

Proof
Suppose that and are similar matrices; that is, for some invertible matrix . Then and for any vector (??) is valid if and only if Thus, when proving this theorem, we may assume that is in Jordan normal form.

Suppose that is in block diagonal form; that is, suppose where is an matrix and is a matrix. Then So for every vector (??) is valid if and only if So, when proving this theorem, we may assume that is a Jordan block.

Consider the case of a simple Jordan block. Suppose that and that where is either real or complex. Then It follows that (??) is valid precisely when . Next, suppose that is a nontrivial Jordan block. For example, let where . It follows by induction that Thus (??) is valid precisely when . The reason for this convergence is as follows. The first term converges to as before but the second term is the product of three terms , , and . The first increases to infinity, the second decreases to zero, and the third is constant independent of . In fact, geometric decay (, when ) always beats polynomial growth. Indeed,

for any integer . This fact can be proved using l’Hôspital’s rule and induction.

So we see that when has a nontrivial Jordan block, convergence is subtler than when has only simple Jordan blocks, as initially the vectors grow in magnitude. For example, suppose that and . Then is the first vector in the sequence whose norm is less than ; that is, is the first vector in the sequence closer to the origin than .

It is also true that (??) is valid for any Jordan block and for all precisely when . To verify this fact we use the binomial theorem. We can write a nontrivial Jordan block as where for some integer . We just discussed the case . In this case

where To verify that we need only verify that each term Such terms are the product of three terms The first term has polynomial growth to infinity dominated by , the second term decreases to zero geometrically, and the third term is constant independent of . The desired convergence to zero follows from (??).

Proof
Recall from Chapter ??, Definition ?? that a Markov matrix is a square matrix whose entries are nonnegative, whose rows sum to , and for which a power that has all positive entries. To prove this theorem we must show that all eigenvalues of satisfy and that is a simple eigenvalue of .

Let be an eigenvalue of and let be an eigenvector corresponding to the eigenvalue . We prove that . Choose so that for all . Since , we can equate the coordinates of both sides of this equality, obtaining Therefore, since the are nonnegative. It follows that since and rows of sum to . Since , it follows that .

Next we show that is a simple eigenvalue of . Recall, or just calculate directly, that the vector is an eigenvector of with eigenvalue . Now let be an eigenvector of with eigenvalue . Let so that all entries of are positive. Observe that is an eigenvector of with eigenvalue , and hence that all rows of also sum to .

To show that is a simple eigenvalue of , and therefore of , we must show that all coordinates of are equal. Using the previous estimates (with ), we obtain

Hence This equality is valid only if all of the are nonnegative or all are nonpositive. Without loss of generality, we assume that all . It follows from (??) that Since , this inequality can hold only if all of the are equal.

Proof
(a) After a similarity transformation, if needed, we can assume that is in Jordan normal form. More precisely, we can assume that where is an matrix with all eigenvalues satisfying . Suppose . It follows from Theorem ?? that Since is the eigenvector of with eigenvalue Part (a) is proved.

(b) Theorem ?? states that a Markov matrix has a dominant eigenvalue equal to . The Jordan normal form theorem implies that the eigenvalues of are equal to the eigenvalues of with the same algebraic and geometric multiplicities. It follows that is also a dominant eigenvalue of . It follows from Part (a) that for some scalar . But Theorem ?? in Chapter ?? implies that the sum of the entries in equals the sum of the entries in which, by assumption equals the sum of the entries in . Thus, .

Exercises

Let be an matrix. Suppose that for every vector . Then the eigenvalues of all satisfy .