Jordan form

It is sound geometric intuition that makes most of us conjecture that, for linear transformations, being invertible and being in some sense zero are exactly opposite notions. Our disappointment in finding that the range and the null-space need not be disjoint is connected with this conjecture. The situation can be straightened out by relaxing the sense in which we interpret "being zero"; for most practical purposes a linear transformation some power of which is zero (that is, a nilpotent transformation) is as zeroish as we can expect it to be. Although we cannot say that a linear transformation is either invertible or "zero" even in the extended sense of zeroness, we can say how any transformation is made up of these two extreme kinds.

Theorem 1. Every linear transformation on a finite-dimensional vector space is the direct sum of a nilpotent transformation and an invertible transformation.

Proof. We consider the null-space of the -th power of ; this is a subspace . Clearly . We assert first that if ever , then for all positive integers . Indeed, if , then , whence (by the fact that ) it follows that , and therefore that . In other words, is contained in (and therefore equal to) ; induction on establishes our assertion.

Since is finite-dimensional, the subspaces cannot continue to increase indefinitely; let be the smallest positive integer for which . It is clear that is invariant under (in fact each is such). We write for the range of (so that, again, it is clear that is invariant under ); we shall prove that and that on is nilpotent, whereas on it is invertible. If is a vector common to and , then and for some . It follows that , and hence, from the definition of , that . We have shown thus that the range and the null-space of are disjoint; a dimensionality argument (see Section: Rank and nullity , Theorem 1) shows that they span , so that is their direct sum. It follows from the definitions of and that on is nilpotent of index . If, finally, is in (so that for some ) and if , then , whence ; this shows that is invertible on . The proof of Theorem 1 is complete. ◻

The decomposition of into its nilpotent and invertible parts is unique. Suppose, indeed, that so that on is nilpotent and on is invertible. Since for some , it follows that , and, since for all , it follows that ; these facts together imply that and .

We can now use our results on nilpotent transformations to study the structure of arbitrary transformations. The method of getting a nilpotent transformation out of an arbitrary one may seem like a conjuring trick, but it is a useful trick, which is often employed. What is essential is the guaranteed existence of proper values; for that reason we continue to assume that the scalar field is algebraically closed (see Section: Multiplicity ).

Theorem 2. If is a linear transformation on a finite-dimensional vector space , and if are the distinct proper values of with respective algebraic multiplicities , then is the direct sum of subspaces of respective dimensions , such that each is invariant under and such that is nilpotent on .

Proof. Take any fixed , and consider the linear transformation . To we may apply the decomposition of Theorem 1 to obtain subspaces and such that is nilpotent on and invertible on . Since is invariant under , it is also invariant under . Hence, for every , the determinant of is the product of the two corresponding determinants for the two linear transformations that becomes when we consider it on and separately. Since the only proper value of on is , and since on does not have the proper value (that is, is invertible on ), it follows that the dimension of is exactly and that each of the subspaces is disjoint from the span of all the others. A dimension argument proves that and thereby concludes the proof of the theorem. ◻

We proceed to describe the principal results of this section and the preceding one in matricial language. If is a linear transformation on a finite-dimensional vector space , then with respect to a suitable basis of , the matrix of has the following form. Every element not on or immediately below the main diagonal vanishes. On the main diagonal there appear the distinct proper values of , each a number of times equal to its algebraic multiplicity. Below any particular proper value there appear only ’s and ’s, and these in the following way: there are chains of ’s followed by a single , with the lengths of the chains decreasing as we read from top to bottom. This matrix is the Jordan form or the classical canonical form of ; we have if and only if the classical canonical forms of and are the same except for the order of the proper values. (Thus, in particular, a linear transformation is diagonable if and only if its classical canonical form is already diagonal, that is, if every chain of ’s has length zero.)

Let us introduce some notation. Let have distinct proper values , with algebraic multiplicities , as before; let the number of chains of ’s under be , and let the lengths of these chains be . The polynomial defined by is called an elementary divisor of of multiplicity belonging to the proper value . An elementary divisor is called simple if its multiplicity is (so that the corresponding chain length is ); we see that a linear transformation is diagonable if and only if its elementary divisors are simple.

To illustrate the power of Theorem 2 we make one application. We may express the fact that the transformation on is nilpotent of index by saying that the transformation on is annulled by the polynomial . It follows that on is annulled by the product of these polynomials (that is, by the product of the elementary divisors of the highest multiplicities); this product is called the minimal polynomial of .

It is quite easy to see (since the index of nilpotence of on is exactly ) that this polynomial is uniquely determined (up to a multiplicative factor) as the polynomial of smallest degree that annuls . Since the characteristic polynomial of is the product of all the elementary divisors, and therefore a multiple of the minimal polynomial, we obtain the Hamilton-Cayley equation : every linear transformation is annulled by its characteristic polynomial.

EXERCISES

Exercise 1. Find the Jordan form of .

Exercise 2. What is the maximum number of pairwise non-similar linear transformations on a three-dimensional vector space, each of which has the characteristic polynomial ?

Exercise 3. Does every invertible linear transformation have a square root? (To say that is a square root of means, of course, that .)

Exercise 4. 

  1. Prove that if is a cube root of ( ), then the matrices are similar.
  2. Discover and prove a generalization of (a) to higher dimensions.

Exercise 5. 

  1. Prove that the matrices are similar.
  2. Discover and prove a generalization of (a) to higher dimensions.

Exercise 6. 

  1. Show that the matrices are similar (over, say, the field of complex numbers).
  2. Discover and prove a generalization of (a) to higher dimensions.

Exercise 7. If two real matrices are similar over , then they are similar over .

Exercise 8. Prove that every matrix is similar to its transpose.

Exercise 9. If and are -by- matrices such that the -by- matrices are similar, then and are similar.

Exercise 10. Which of the following matrices are diagonable (over the field of complex numbers)?

  1. ,
  2. ,
  3. ,
  4. ,
  5. .

What about the field of real numbers?

Exercise 11. Show that the matrix is diagonable over the field of complex numbers but not over the field of real numbers.

Exercise 12. Let be a permutation of the integers ; if is a vector in , write . Prove that is diagonable and find a basis with respect to which the matrix of is diagonal.

Exercise 13. Suppose that is a linear transformation and that is a subspace invariant under . Prove that if is diagonable, then so also is the restriction of to .

Exercise 14. Under what conditions on the complex numbers is the matrix diagonable (over the field of complex numbers)?

Exercise 15. Are the following assertions true or false?

  1. A real two-by-two matrix with a negative determinant is similar to a diagonal matrix.
  2. If is a linear transformation on a complex vector space, and if for some positive integer , then is diagonable.
  3. If is a nilpotent linear transformation on a finite-dimensional vector space, then is diagonable.

Exercise 16. If is a linear transformation on a finite-dimensional vector space over an algebraically closed field, and if every proper value of has algebraic multiplicity , then is diagonable.

Exercise 17. If the minimal polynomial of a linear transformation on an -dimensional vector space has degree , then is diagonable.

Exercise 18. Find the minimal polynomials of all projections and all involutions.

Exercise 19. What is the minimal polynomial of the matrix

Exercise 20. 

  1. What is the minimal polynomial of the differentiation operator on ?
  2. What is the minimal polynomial of the transformation on defined by ?

Exercise 21. If is a linear transformation with minimal polynomial , and if is a polynomial such that , then is divisible by .

Exercise 22. 

  1. If and are linear transformations, if is a polynomial such that , and if , then .
  2. What can be inferred from (a) about the relation between the minimal polynomials of and of ?

Exercise 23. A linear transformation is invertible if and only if the constant term of its minimal polynomial is different from zero.