Uncorrelated and Independent Random Variables

The notion of independence of two random variables, and , is defined in section 6 of Chapter 7 . In this section we show how the notion of independence may be formulated in terms of expectations. At the same time, by a modification of the condition for independence of random variables, we are led to the notion of uncorrelated random variables.

We begin by considering the properties of expectations of products of random variables. Let and be jointly distributed random variables. By the linearity properties of the operation of taking expectations, it follows that for any two functions, and if the expectations on the right side of (3.1) exist. However, it is not true that a similar relation holds for products; namely, it is not true in general that . There is one special circumstance in which a relation similar to the foregoing is valid, namely, if the random variables and are independent and if the functions are functions of one variable only. More precisely, we have the following theorem:

Theorem 3A: If the random variables and are independent, then for any two Borel functions and of one real variable the product moment of and is equal to the product of their means; in symbols, 

if the expectations on the right side of (3.2) exist. 

To prove equation (3.2), it suffices to prove it in the form

since independence of and implies independence of and . We write out the proof of (3.3) only for the case of jointly continuous random variables. We have

Now suppose that we modify (3.2) and ask only that it hold for the functions and , so that

For reasons that are explained after (3.7), two random variables, and , which satisfy (3.4), are said to be uncorrelated. From (2.10) it follows that and satisfy (3.4) and therefore are uncorrelated if and only if

For uncorrelated random variables the formula given by (2.11) for the variance of the sum of two random variables becomes particularly elegant; the variance of the sum of two uncorrelated random variables is equal to the sum of their variances. Indeed, 

if and only if and are uncorrelated. 

Two random variables that are independent are uncorrelated, for if (3.2) holds then, a fortiori, (3.4) holds. The converse is not true in general; an example of two uncorrelated random variables that are not independent is given in theoretical exercise 3.2. In the important special case in which and are jointly normally distributed, it follows that they are independent if they are uncorrelated (see theoretical exercise 3.3).

The correlation coefficient of two jointly distributed random variables with finite positive variances is defined by

In view of (3.7) and (3.5), two random variables and are uncorrelated if and only if their correlation coefficient is zero.

The correlation coefficient provides a measure of how good a prediction of the value of one of the random variables can be formed on the basis of an observed value of the other. It is subsequently shown that

Further if and only if

and if and only if

From (3.9) and (3.10) it follows that if the correlation coefficient equals 1 or -1 then there is perfect prediction; to a given value of one of the random variables there is one and only one value that the other random variable can assume. What is even more striking is that if and only if and are linearly dependent.

That (3.8), (3.9), and (3.10) hold follows from the following important theorem.

Theorem 3B . For any two jointly distributed random variables, and , with finite second moments Further, equality holds in (3.11), that is, if and only if, for some constant , which means that the probability mass distributed over the -plane by the joint probability law of the random variables is situated on the line

Applied to the random variables and , (3.11) states that

We prove (3.11) as follows. Define, for any real number . Clearly for all . Consequently, the quadratic equation has either no solutions or one solution. The equation has no solutions if and only if . It has exactly one solution if and only if . From these facts one may immediately infer (3.11) and the sentence following it.

The inequalities given by (3.11) and (3.12) are usually referred to as Schwarz’s inequality or Cauchy’s inequality.

Conditions for Independence . It is important to note the difference between two random variables being independent and being uncorrelated. They are uncorrelated if and only if (3.4) holds. It may be shown that they are independent if and only if (3.2) holds for all functions and , for which the expectations in (3.2) exist. More generally, theorem can be proved.

Theorem 3c. Two jointly distributed random variables and are independent if and only if each of the following equivalent statements is true: 

(i) Criterion in terms of probability functions. For any Borel sets and of real numbers, is in is in is in is in

(ii) Criterion in terms of distribution functions. For any two real numbers, and

(iii) Criterion in terms of expectations. For any two Borel functions, and if the expectations involved exist. 

(iv) Criterion in terms of moment-generating functions (if they exist). For any two real numbers, and

Theoretical Exercises

3.1. The standard deviation has the properties of the operation of taking the absolute value of a number : show first that for any 2 real numbers, and .

Hint : Square both sides of the equations. Show next that for any 2 random variables, and ,

Give an example to prove that the variance does not satisfy similar relationships.

3.2. Show that independent random variables are uncorrelated. Give an example to show that the converse is false.

Hint : Let , , in which is uniformly distributed over the interval 0 to 1.

3.3. Prove that if and are jointly normally distributed random variables whose correlation coefficient vanishes then and are independent. Hint : Use example 2A .

3.4 . Let and be the values of and which minimize

Express , and in terms of . The random variable is called the best linear predictor of , given [see Section 7, in particular, (7.13) and (7.14)].

3.5. Prove that (3.9) and (3.10) hold under the conditions stated.

3.6. Let and be jointly distributed random variables possessing finite second moments. State conditions under which it is possible to find 2 uncorrelated random variables, and , which are linear combinations of and (that is, and for some constants and ).

3.7. Let and be jointly normally distributed with mean 0, arbitrary variances, and correlation . Show that Hint : Consult H. Cramér, Mathematical Methods of Statistics , Princeton University Press, 1946, p. 290.

3.8. Suppose that tickets bear arbitrary numbers , which are not all the same. Suppose further that 2 of the tickets are selected at random without replacement. Show that the correlation coefficient between the numbers appearing on the 2 tickets is equal to .

3.9. In an urn containing balls, a proportion is white and are black. A ball is drawn and its color noted. The ball drawn is then replaced, and balls are added of the same color as the ball drawn. The process is repeated until balls have been drawn. For let be equal to 1 or 0, depending on whether the ball drawn on the th draw is white or black. Show that the correlation coefficient between and is equal to . Note that the case corresponds to sampling without replacement, and corresponds to sampling with replacement.

Exercises

3.1. Consider 2 events and such that . Define random variables and or 0, depending on whether the event has or has not occurred, and or 0, depending on whether the event has or has not occurred. Find , . Are and independent?

 

Answer

and are independent.

 

3.2. Consider a sample of size 2 drawn with replacement (without replacement) from an urn containing 4 balls, numbered 1 to 4. Let be the smallest and be the largest among the numbers drawn in the sample. Find .

3.3. Two fair coins, each with faces numbered 1 and 2, are thrown independently. Let denote the sum of the 2 numbers obtained, and let denote the maximum of the numbers obtained. Find the correlation coefficient between and .

 

Answer

.

 

3.4. Let , and be uncorrelated random variables with equal variances. Let . Find the correlation coefficient between and .

3.5. Let and be uncorrelated random variables. Find the correlation between the random variables and in terms of the variances of and .

 

Answer

.

 

3.6. Let and be uncorrelated normally distributed random variables. Find the correlation between the random variables and .

3.7. Consider the random variables whose joint moment-generating function is given in exercise 2.6 . Find .

 

Answer

.

 

3.8. Consider the random variables whose joint moment-generating function is given in exercise 2.7 . Find .

3.9. Consider the random variables whose joint moment-generating function is given in exercise 2.8 . Find .

 

Answer

.

 

3.10. Consider the random variables whose joint moment-generating function is given in exercise 2.9 . Find .