Change of basis

Although what we have been doing with linear transformations so far may have been complicated, it was to a large extent automatic. Having introduced the new concept of linear transformation, we merely let some of the preceding concepts suggest ways in which they are connected with linear transformations. We now begin the proper study of linear transformations. As a first application of the theory we shall solve the problems arising from a change of basis. These problems can be formulated without mentioning linear transformations, but their solution is most effectively given in terms of linear transformations.

Let be an -dimensional vector space and let and be two bases in . We may ask the following two questions.

Question I. If is in , , what is the relation between its coordinates with respect to and its coordinates with respect to ?

Question II. If is an ordered set of scalars, what is the relation between the vectors and ?

Both these questions are easily answered in the language of linear transformations. We consider, namely, the linear transformation defined by , . More explicitly: Let be the matrix of in the basis , that is, . We observe that is invertible, since implies that .

Answer to question I. Since we have

Answer to question II.

Roughly speaking, the invertible linear transformation (or, more properly, the matrix may be considered as a transformation of coordinates (as in (1)), or it may be considered (as we usually consider it, in (2)) as a transformation of vectors.

In classical treatises on vector spaces it is customary to treat vectors as numerical -tuples, rather than as abstract entities; this necessitates the introduction of some cumbersome terminology. We give here a brief glossary of some of the more baffling terms and notations that arise in connection with dual spaces and adjoint transformations.

If is an -dimensional vector space, a vector is given by its coordinates with respect to some preferred, absolute coordinate system; these coordinates form an ordered set of scalars. It is customary to write this set of scalars in a column, Elements of the dual space are written as rows, . If we think of as a (rectangular) -by-one matrix, and of as a one-by- matrix, then the matrix product is a one-by-one matrix, that is, a scalar. In our notation this scalar is . The trick of considering vectors as thin matrices works even when we consider the full-grown matrices of linear transformations. Thus the matrix product of with the column is the column whose -th element is . Instead of worrying about dual bases and adjoint transformations, we may form similarly the product of the row with the matrix in the order ; the result is the row that we earlier denoted by . The expression is now abbreviated as ; both dots denote ordinary matrix multiplication. The vectors in are called covariant and the vectors in are called contravariant . Since the notion of the product (that is, ) depends, from this point of view, on the coordinates of and , it becomes relevant to ask the following question: if we change basis in , in accordance with the invertible linear transformation , what must we do in to preserve the product ? In our notation: if , where , then how is related to ? Answer: . To express this whole tangle of ideas the classical terminology says that the vectors vary cogrediently whereas the vary contragrediently .