The Central Limit Theorem

A sequence of jointly distributed random variables with finite means and variances is said to obey the (classical) central limit theorem if the sequence , defined by

converges in distribution to a random variable that is normally distributed with mean 0 and variance 1. In terms of characteristic functions, the sequence obeys the central limit theorem if for every real number

The random variables are called the sequence of normalized consecutive sums of the sequence .

That the central limit theorem is true under fairly unrestrictive conditions on the random variables was already surmised by Laplace and Gauss in the early 1800’s. However, the first satisfactory conditions, backed by a rigorous proof, for the validity of the central limit theorem were given by Lyapunov in 1901. In the 1920’s and 1930’s the method of characteristic functions was used to extend the theorem in several directions and to obtain fairly unrestrictive necessary and sufficient conditions for its validity in the case in which the random variables are independent. More recent years have seen extensive work on extending the central limit theorem to the case of dependent random variables.

The reader is referred to the treatises of B. V. Gnedenko and A. N. Kolmogorov, Limit Distributions for Sums of Independent Random Variables , Addison-Wesley, Cambridge, Mass., 1954, and M. Loève, Probability Theory , Van Nostrand, New York, 1955, for a definitive treatment of the central limit theorem and its extensions.

From the point of view of the applications of probability theory, there are two main versions of the central limit theorem that one should have at his command. One should know conditions for the validity of the central limit theorem in the cases in which (i) the random variables are independent and identically distributed and (ii) the random variables are independent but not identically distributed.

Theorem 4A. THE CENTRAL LIMIT THEOREM FOR INDEPENDENT IDENTICALLY DISTRIBUTED RANDOM VARIABLES WITH FINITE MEANS AND VARIANCES . For let be identically distributed as the random variable , with finite mean and standard deviation . Let the sequence be independent, and let be defined by (4.1) or, more explicitly, 

Then (4.2) will hold. 

Theorem 4B. THE CENTRAL LIMIT THEOREM FOR INDEPENDENT RANDOM VARIABLES WITH FINITE MEANS AND th CENTRAL MOMENT, FOR SOME . For let be a random variable with finite mean and finite th central moment

Let the sequence be independent, and let be defined by (4.1). Then (4.2) will hold if 

in which

Equation (4.4) is called Lyapunov’s condition for the validity of the central limit theorem for independent random variables .

We turn now to the proofs of theorems 4A and 4B. Consider first independent random variables , identically distributed as the random variable , with mean 0 and variance . Let be their normalized sum, given by (4.1). The characteristic function of may be written

Now tends to as tends to . Therefore, for each fixed exists (by lemma 3A) when . For as large as this, using the expansion given by (3.8),

Theorem 4A will be proved if we prove that

It is clear that to prove (4.7) will hold we need prove only that the integral in (4.6) tends to 0 as tends to infinity. Define . Then, for any

Now, for and for , in view of the inequalities . From these facts we may conclude that for any and real numbers and

Then

which tends to 0, as we let first tend to and then tend to . The proof of the central limit theorem for identically distributed independent random variables with finite variances is complete.

We next prove the central limit theorem under Lyapunov’s condition. For , let be a random variable with mean 0, finite variance , and th central moment . We have the following expansion of the logarithm of its characteristic function, for such that :

To prove (4.11), merely use in (3.8) the inequality , valid for any real number and .

Now, (4.4) and theoretical exercise 4.3 imply that

Then, for any fixed it holds for sufficiently large that for all . Therefore, exists and is given by

The first sum in (4.13) is equal to 1, whereas the second sum tends to 0 by Lyapunov’s condition, as does the third sum, since

The proof of the central limit theorem under Lyapunov’s condition is complete.

Theoretical exercises

4.1. Prove that the central limit theorem holds for independent random variables with zero means and finite variances obeying Lindeberg’s condition: for every

Hint: In (4.8) let , replacing by . Obtain thereby an estimate for . Add these estimates to obtain an estimate for , as in (4.13).

4.2. Prove the law of large numbers under Markov’s condition. Hint: Adapt the proof of the central limit theorem under Lyapunov’s condition, using the expansions (3.13).

4.3. Jensen’s inequality and its consequences. Let be a tandom variable, and let be a (possibly infinite) interval such that, with probability one, takes its values in ; that is lies in . Let be a function of a real variable that is twice differentiable on and whose second derivative satisfies for all in . The function is then said to be convex on . Show that the following inequality (Jensen’s inequality) holds:

Hint: Show by Taylor’s theorem that . Let and take the expectation of both sides of the inequality. Deduce from (4.15) that for any and

Conclude from (4.17) that if then

In particular, conclude that

4.4. Let be a sequence of independent random variables, each uniformly distributed on the interval 0 to . Let be a sequence of positive constants. State conditions under which the sequence obeys the central limit theorem.