As the sample size n increases, what happens to the standard error of the mean

Example: Mean NFL Salary

The built-in dataset "NFL Contracts (2015 in millions)" was used to construct the two sampling distributions below. In the first, a sample size of 10 was used. In the second, a sample size of 100 was used.

Sample size of 10:

200Show Data TableGenerate 100 Samples Choose samples of size n =samples = 5000mean = 2.195std. error = 0.936Generate 10 SamplesGenerate 1 SampleLeft TailTw o - TailRight TailNFL Contr acts (2015 in millions)Sampling Dotplot of MeanEdit DataUpload File Change ColuReset Plot10150 100500122.1953456789Generate 100 0 Samples

Sample size of 100:

120Show Data TableGenerate 100 SamplesChoose samples of size n =samples = 5000mean = 2.236std. error = 0.296Generate 10 SamplesGenerate 1 SampleLeft TailTwo - TailRight TailNFL Contracts (2015 in millions)Sampling Dotplot of MeanEdit D ataUpload FileChange ColuReset Plot1001008060402001.501.752.002.252.502.753.003.25Generate 1000 Samples2.236

With a sample size of 10, the standard error of the mean was 0.936. With a sample size of 100 the standard error of the mean was 0.296. When the sample size increased the standard error decreased.

Also know that the population was strongly skewed to the right. With the smaller sample size, the sampling distribution was also skewed to the right, though not as strongly skewed as the population. With the larger sample size, the sampling distribution was approximately normal.

  1. Last updated
  2. Save as PDF
  • Page ID4584
  • Examples of the Central Limit Theorem

    Law of Large Numbers

    The law of large numbers says that if you take samples of larger and larger size from any population, then the mean of the sampling distribution, \(\mu_{\overline x}\) tends to get closer and closer to the true population mean, \(\mu\). From the Central Limit Theorem, we know that as \(n\) gets larger and larger, the sample means follow a normal distribution. The larger n gets, the smaller the standard deviation of the sampling distribution gets. (Remember that the standard deviation for the sampling distribution of \(\overline X\) is \(\frac{\sigma}{\sqrt{n}}\).) This means that the sample mean \(\overline x\) must be closer to the population mean \(\mu\) as \(n\) increases. We can say that \(\mu\) is the value that the sample means approach as n gets larger. The Central Limit Theorem illustrates the law of large numbers.

    This concept is so important and plays such a critical role in what follows it deserves to be developed further. Indeed, there are two critical issues that flow from the Central Limit Theorem and the application of the Law of Large numbers to it. These are

    1. The probability density function of the sampling distribution of means is normally distributed regardless of the underlying distribution of the population observations and
    2. standard deviation of the sampling distribution decreases as the size of the samples that were used to calculate the means for the sampling distribution increases.

    Taking these in order. It would seem counterintuitive that the population may have any distribution and the distribution of means coming from it would be normally distributed. With the use of computers, experiments can be simulated that show the process by which the sampling distribution changes as the sample size is increased. These simulations show visually the results of the mathematical proof of the Central Limit Theorem.

    Here are three examples of very different population distributions and the evolution of the sampling distribution to a normal distribution as the sample size increases. The top panel in these cases represents the histogram for the original data. The three panels show the histograms for 1,000 randomly drawn samples for different sample sizes: \(n=10\), \(n= 25\) and \(n=50\). As the sample size increases, and the number of samples taken remains constant, the distribution of the 1,000 sample means becomes closer to the smooth line that represents the normal distribution.

    Figure \(\PageIndex{3}\) is for a normal distribution of individual observations and we would expect the sampling distribution to converge on the normal quickly. The results show this and show that even at a very small sample size the distribution is close to the normal distribution.

    Figure \(\PageIndex{3}\)

    Figure \(\PageIndex{4}\) is a uniform distribution which, a bit amazingly, quickly approached the normal distribution even with only a sample of 10.

    Figure \(\PageIndex{4}\)

    Figure \(\PageIndex{5}\) is a skewed distribution. This last one could be an exponential, geometric, or binomial with a small probability of success creating the skew in the distribution. For skewed distributions our intuition would say that this will take larger sample sizes to move to a normal distribution and indeed that is what we observe from the simulation. Nevertheless, at a sample size of 50, not considered a very large sample, the distribution of sample means has very decidedly gained the shape of the normal distribution.

    Figure \(\PageIndex{5}\)

    The Central Limit Theorem provides more than the proof that the sampling distribution of means is normally distributed. It also provides us with the mean and standard deviation of this distribution. Further, as discussed above, the expected value of the mean, \(\mu_{\overline{x}}\), is equal to the mean of the population of the original data which is what we are interested in estimating from the sample we took. We have already inserted this conclusion of the Central Limit Theorem into the formula we use for standardizing from the sampling distribution to the standard normal distribution. And finally, the Central Limit Theorem has also provided the standard deviation of the sampling distribution, \(\sigma_{\overline{x}}=\frac{\sigma}{\sqrt{n}}\), and this is critical to have to calculate probabilities of values of the new random variable, \(\overline x\).

    Figure \(\PageIndex{6}\) shows a sampling distribution. The mean has been marked on the horizontal axis of the \(\overline X\)'s and the standard deviation has been written to the right above the distribution. Notice that the standard deviation of the sampling distribution is the original standard deviation of the population, divided by the sample size. We have already seen that as the sample size increases the sampling distribution becomes closer and closer to the normal distribution. As this happens, the standard deviation of the sampling distribution changes in another way; the standard deviation decreases as \(n\) increases. At very very large \(n\), the standard deviation of the sampling distribution becomes very small and at infinity it collapses on top of the population mean. This is what it means that the expected value of \(\mu_{\overline{x}}\) is the population mean, \(\mu\).

    Figure \(\PageIndex{6}\)

    At non-extreme values of \(n\), this relationship between the standard deviation of the sampling distribution and the sample size plays a very important part in our ability to estimate the parameters we are interested in.

    Figure \(\PageIndex{7}\) shows three sampling distributions. The only change that was made is the sample size that was used to get the sample means for each distribution. As the sample size increases, \(n\) goes from 10 to 30 to 50, the standard deviations of the respective sampling distributions decrease because the sample size is in the denominator of the standard deviations of the sampling distributions.

    Figure \(\PageIndex{7}\)

    The implications for this are very important. Figure \(\PageIndex{8}\) shows the effect of the sample size on the confidence we will have in our estimates. These are two sampling distributions from the same population. One sampling distribution was created with samples of size 10 and the other with samples of size 50. All other things constant, the sampling distribution with sample size 50 has a smaller standard deviation that causes the graph to be higher and narrower. The important effect of this is that for the same probability of one standard deviation from the mean, this distribution covers much less of a range of possible values than the other distribution. One standard deviation is marked on the \(\overline X\) axis for each distribution. This is shown by the two arrows that are plus or minus one standard deviation for each distribution. If the probability that the true mean is one standard deviation away from the mean, then for the sampling distribution with the smaller sample size, the possible range of values is much greater. A simple question is, would you rather have a sample mean from the narrow, tight distribution, or the flat, wide distribution as the estimate of the population mean? Your answer tells us why people intuitively will always choose data from a large sample rather than a small sample. The sample mean they are getting is coming from a more compact distribution. This concept will be the foundation for what will be called level of confidence in the next unit.

    Figure \(\PageIndex{8}\)

    What happens to the standard error of mean as sample size increases?

    Standard error decreases when sample size increases – as the sample size gets closer to the true size of the population, the sample means cluster more and more around the true population mean.

    How does sample size n affect standard error?

    From the formula, you'll see that the sample size is inversely proportional to the standard error. This means that the larger the sample, the smaller the standard error, because the sample statistic will be closer to approaching the population parameter.

    What happens when sample size n increases?

    As the sample sizes increase, the variability of each sampling distribution decreases so that they become increasingly more leptokurtic. The range of the sampling distribution is smaller than the range of the original population.

    What happens to the standard error when n is very large?

    Larger sample sizes have narrower sampling distributions. Since n is in the denominator of the Standard Error formula, as n increases Standard Error decreases.

    Toplist

    Neuester Beitrag

    Stichworte