We can now prove one of the essential results in probability: the chance that a sample average differs from its expectation by any fixed threshold converges to zero as the number of samples diverges, no matter how small the threshold, as long as the samples are sufficiently uncorrelated. This result is the weak law of large numbers.
In particular, if the samples are drawn independently and identically then μn=E[Xj]=μ and:
You can experiment with the weak law of large numbers using this Law of Large Numbers Interactive. You can choose the distribution that produces X, then track Xˉn as n increases. You can also choose the tolerance (window half-width) about the true expected value. You will see that, as n increases, the sample average eventually settles within the tolerance of its expectation. The panel on the right shows the empirical distribution of the sampled values {Xj}j=1n. As n increases, the empirical distribution converges to the underlying distribution, so its expectation (the sample average) converges to the underlying expected value.
In Section 1.2 we defined chances as long run frequencies. The law of large numbers makes this definition more concrete.
Let E be some event. Let Ij be an indicator for the event on the jth trial of a sequence of n independent and identical repetitions of the random process. Then, the frequency of the event over n trials is the sample average of the indicators:
The variance in an indicator random variable is p(1−p)≤1/4. So, the probability that an empirical frequency differs from its expectation can be bounded using Chebyshev’s inequality. The expected value of an indicator is the success probability for the corresponding event so:
In other words, the probability that the observed frequency with which an event occurs in a sequence of independent, identical repetitions, differs from the chance of the event by more than some tolerance, converges to zero at rate O(1/n) or faster, no matter the tolerance. Thus, observed frequencies converge to underlying chances!