Point estimates are random variables. Random variables follow shapes, called distributions. Therefore, point estimates follow distributions (and have shapes).
The Central Limit Theorem says, “If our sample size is large enough, the sample mean will be approximately Normally distributed.”
The Central Limit Theorem says, if we have a collection of sample means, the shape (histogram) of this collection is basically Normal (unimodal and symmetric).
The Central Limit Theorem says, “Under certain conditions, the sampling distribution for the sample mean converges to the normal distribution as the sample size increases.”
When \(n\) is sufficiently large,
\[\frac{\bar{X} - \mu}{\sigma_{\bar{X}}} \overset{\cdot}{\sim} N(0, 1).\]
As \(n\) increases, the approximation improves.
We should mention the conditions necessary for this to happen.
The assumptions are really not that bad, so we can safely assume they hold in many real world applications. With that, so long as we use the mean, then we can say, at least approximately, how the distribution of means is shaped – even if we never actually sample/calculate multiple means.
It doesn’t matter if that data are from a Normal distribution or not; use mean => Central Limit Theorem.
We will use a sample of Darwin’s finch data set [Swarth:1931]. Make a 98% confidence interval about the beak height using Darwin’s finches.
98% confidence interval for Darwin’s finch data set.
\[\bar{x} \pm t * \frac{std}{\sqrt{n}}\]
Because math/statistics