Limits of Linear Transformations


 

1. Convergence of vectors

In this chapter we shall indicate the various possibilities for defining the concept of limit for sets of vectors or linear transformations, and we shall give a few of the many possible applications of these concepts.

If is an -dimensional inner product space, the two fundamental quantitive concepts—inner product, and length—suggest the following two definitions of convergence in .

Definition 1. We say that a sequence, , of vectors converges to the vector in the strong sense if as ; and in the weak sense if for every .

It turns out that this distinction is merely pedantic in our finite dimensional case: the two notions of convergence are equivalent. For if (i.e., converges strongly to ) then Conversely if , and if is an orthogonal basis in , , , then converges, as , to , for each . Consequently

as , as was to be proved.

Concerning the convergence (in either of the two equivalent senses) of vectors, we shall use without proof the following facts. The expression is a continuous function of all its arguments simultaneously, i.e., if and are sequences of complex numbers and and vectors, then implies that

If is an orthogonal basis in , , , then a necessary and sufficient condition for the convergence of to is that for each . (Thus the notion of convergence here defined coincides with the usual one in -dimensional complex Euclidean space.) Finally we shall assume as known the fact that the metric topology defined by the notion of convergence is complete: i.e., if is a sequence of vectors for which as , then there is a unique vector such that as .

2. Convergence of linear transformations

The notion of convergence for linear transformations is apparently more complicated than that for vectors. Using the ideas of the preceding section, and remembering our discussion of the norm, , of a linear transformation , we are led to consider the following three definitions for the convergence of a sequence of linear transformations to the linear transformation .

Definition 2. in the uniform sense if ; in the strong sense if in the strong sense for every vector ; and in the weak sense if in the weak sense.

Once more the distinctions are illusory: we shall prove that the three notions of convergence coincide. That weak and strong convergence are equivalent follows already from the results of the preceding section. If , then for every we have

so that uniform convergence implies strong convergence. To prove the converse, let be an orthogonal basis in , let be an arbitrary vector, and suppose that . Then for any we may find an such that for and we have . It follows that so that for , . This completes the proof of the equivalence of the three concepts, and yields also an easy proof of the fact that again our topology is complete. If as , then for each , . Hence to each there corresponds a vector, which we may designate by , such that . It is easy to verify that the correspondence from to is given by a linear transformation and that in the strong sense. Consequently in the uniform sense, and this proves our assertion concerning completeness.

If is an orthogonal basis in which the matrices of and are and respectively, then a necessary and sufficient condition for the convergence of (in any one of the three equivalent senses given above) is that as for each . This follows from the fact that and the definition of weak convergence. Consequently the concepts of convergence we discussed reduce to the usual concept in dimensional complex or dimensional real Euclidean space. It would be easy to use this fact in the proof of several statements given below: we prefer a slightly more circuitous route, which sheds more light on the concept of the norm of a transformation.

3. Inequalities on norms

In this section we fill a gap in our discussion of norms: our purpose is to prove the following four relations:

All of the proofs are easy from the definition: we sketch the proofs for the product and the dual. We have

For duals we have It follows that ; by symmetry we obtain the desired result.

4. The continuity of the ring operations

is a continuous function of ; is a continuous function of ; , , , and are continuous functions of all their arguments. Of these statements we shall prove only the ones concerning , , and : the others are proved by very similar methods.

4.1. From the relations and we conclude that the absolute value of the difference is dominated by . Since means , we see that the convergence of to implies .

In particular we note that implies that remains bounded.

4.2. If , , then

Since and are bounded and the other factors converge to zero, we see that .

4.3. If then for each and , converges to . Hence converges to in the weak and therefore in the uniform sense.

We shall now prove a few theorems about limits of special linear transformations by way of illustration of the general theory.

5. The mean ergodic theorem

Theorem 1. If is a unitary transformation and is a subspace of all vectors for which , then

where is the projection on .

Proof. For let be the spectral form of ; we have and . Moreover is equal to one of the : that one for which . (we do not exclude the possibility that no , so that .) Hence we have and it is sufficient to prove that the numbers converges to or according as or . In case this is obvious; In any other case we have

6. Markoff matrices

To give our next application of the notion of convergence we forsake the theory of linear transformations and consider matrices, and we shall understand by the convergence of a sequence of matrices to the numercial convergence for each , .

Definition 3. A Markoff matrix is a matrix for which and .

Concerning Markoff matrices we need the following two lemmas.

Lemma 1. The eigenvalues of a Markoff matrix are in absolute value .

Proof. Since the eigenvalues of and of are the same, it is sufficient to prove that implies . We have and therefore Since , we have and from this the desired conclusion is immediate.

Lemma 2. The product of two Markoff matrices and , is a Markoff matrix.

Proof. Write ; it is clear that . Also

so that is a Markoff matrix.

As a consequence of this lemma we obtain the fact that the totality of all elements in the infinite sequence is bounded, and therefore that the same is true of the sequence of powers of , where is an arbitrary non-singular matrix. The last statement follow from the fact that the elements of are linear combinations, with coefficients which are elements of and of , of the elements of .

7. Limits of Markoff matrices

Our main theorem concerning Markoff matrices is the following:

Theorem 2. For an arbitrary Markoff matrix , converges to a limit as ; a necessary and sufficient condition that not only the averages, but also the powers themselves converge to a limit is that be the only eigenvalue of of absolute value one.

Proof. To prove this result it is sufficient to prove it not for but for , with a suitably chosen . Since we may choose so that is in the classical (Jordan) canonical form, we may restrict our attention to matrices of the form , where

In this case we have Hence, the general non-vanishing element in is a number of the form , where is a fixed positive integer depending only on the place of the element and independent of .

Since is an eigenvalue of a Markoff matrix we know that . If we have for a suitable constant , and therefore in this case not only the averages of the elements of , but even the elements themselves converge to zero.

If, however, , then , so that in this case the elements of do not remain bounded. Since is a transform of a Markoff matrix, this is impossible: the only possible conclusion is that the disturbing super diagonal ’s do not enter in this case, so that the boxes of the classical canonical form which correspond to eigenvalues of absolute value one are of size . In this case therefore the problem is reduced to the convergence properties of , where : we have already seen that the averages of this sequence always converge, to zero or one according as or , and that the sequence itself converges if and only if .

This completes the proof of all our assertions: we see even that since the limit, after transforming by , is a diagonal matrix whose diagonal elements consist only of zeros and ones, the limit is in general idempotent.

8. The exponential function

Our final example of limits is furnished by , where is an arbitrary linear transformation in the inner product space.

Consider the transformations ; let be the norm, , of the transformation . Since we have

and since the dominant on the right, being a part of the power series for , converges to zero as , it follows that the series converges, or, in other words that the have a limit. This limit we define to be : we shall mention some of the elementary properties of this exponential function.

Since we may assume that is a matrix in upper triangular form, so that the diagonal elements are the eigenvalues of , each occurring with its proper multiplicity, it follows that the eigenvalues of are the exponentials of those of . From the consideration of the upper triangular form it follows also that the determinant of , i.e., , where are the diagonal elements of the upper triangular form of , is the same as , i.e., the exponential of the trace of . This shows incidentally that is never a singular transformation.

We conclude by studying the multiplication rule for exponentials. We suppose that and are commutative transformations. Since is the limit of the expression we will have proved the multiplication rule when we have proved that this expression converges to zero as . An easy verification yields the fact that for , occur in both members with equal coefficients; the terms that do not cancel out are all in the right member and are equal to the summation being extended over those values of for which . implies that at least one of and is greater than the integer part of , the bound of this remainder is dominated by

where and as .