Often, when one wants to solve various scientific problems, several assumptions will be made regarding the nature of the underlying setting and base their conclusions on those assumptions. Indeed, if one is going to use a Binomial Distribution or a Negative Binomial Distribution, an assumption on the value of p is necessary. For Poisson and Exponential Distributions, one must know the mean. For Normal Distributions, one must assume values for both the mean and the standard deviation. Where do these values come from? Often, one may perform a preliminary study and obtain a sample statistic...such as a sample mean or a relative frequency and use these values for ΞΌ or p.
Often the Central Limit Theorem is stated more formally using a conversion to standard units. Indeed, the theorem indicates that the random variable
has variance
which means as n grows this variance approaches 0. So, the limiting random variable has a zero variance and therefore is no longer a random variable. To avoid this issue, the Central Limit Theorem is often stated as:
When converting probability problems from continuous (such as exponential or uniform) then no adjustment to the question is needed since you are approximating one area with another area. However, when converting probability problems from discrete (such as binomial or geometric) then you need to consider how the interval would need to be adjusted so that histogram areas for the discrete problem would relate to areas under the normal curve. Generally, you will need to expand the stated interval each way by 1/2.
The Central Limit Theorem provides that regardless of the distribution of X, the distribution of an average of Xβs is approximately normally distributed. However, it also shows why X may also be approximated for some distributions using the normal distribution as certain parameters are allowed to increase. Below, you can see how Binomial and Poisson distributions can be approximated directly using the Normal distribution.
Binomial becomes normal as Consider n = 50 and p = 0.3. Then, and
Using the binomial formulas, for example,
Using the normal distribution,
Notice that these are very close.
Poisson becomes normal as Consider Then,
Using the Poisson formulas, for example,
Using the normal distribution,
Again, these are very close.
Gamma becomes normal as Assume that the average time till a first success is 12 minutes and that Then, the mean for the Gamma distribution is and and so
Using the Gamma formulas,
Using the normal distribution,
Amazingly, these are also very close.
Consider a discrete uniform variable X over R = {1,2,...,20}. Then, and using the uniform formulas.
You can use the uniform probability function to compute probabilities dealing with X. Indeed,
If instead you plan to sample from this distribution n=49 times, the Central Limit Theorem implies that you will get a random variable which has an approximate normal distribution with the same mean but with new variance Therefore, expanding the interval to include the boundaries of the corresponding histogram areas,
As these examples illustrate, you will have increasing success in approximating the desired probabilities so long as the distributionβs corresponding parameter is allowed to be "sufficiently large". The mathematical reasoning this is true is not provided but depends upon the "Central Limit Theorem" discussed in the next section.
The above theorems allow you to utilize the normal distribution to compute approximate probabilities for the variable X in the stated distributions. This is not always true for all distributions since some do not have parameters which allow for approaching normality. However, regardless of the distribution the Central Limit Theorem always allows you to approximate probabilities if they involve an average of repeated attempts...that is, for variable
This usefulness is illustrated in the examples below.