Normal, t, Chi-Square, and F Distributions
Several continuous distributions appear repeatedly in introductory inference. The normal distribution models many measurements and approximates many sampling distributions. The distribution appears when estimating a mean with an unknown population standard deviation. The chi-square distribution appears in variance inference and categorical-data tests. The distribution appears in ANOVA and regression model comparisons. The Lane text treats these distributions across several chapters, but they form one connected family in practice.
These distributions are not just named curves. Each one describes the behavior of a particular statistic under assumptions. Understanding the statistic behind the curve prevents mechanical table lookup. For example, a statistic compares an estimated effect to its estimated standard error, while an statistic compares variance explained by a model to variance left unexplained.
Definitions
A normal distribution with mean and standard deviation is written
Its density is bell-shaped, symmetric around , and controlled by . The standard normal distribution is . Standardization converts a normal variable to standard normal:
If is standard normal, probabilities are areas under the standard normal curve.
The distribution with degrees of freedom is symmetric around 0 and has heavier tails than the standard normal distribution. It is commonly used for
when sampling from a normal population and estimating by . The degrees of freedom for a one-sample statistic are .
The chi-square distribution with degrees of freedom is written . If are independent standard normal variables, then
has a chi-square distribution with degrees of freedom. The distribution is nonnegative and right-skewed, especially for small .
The distribution has two degrees of freedom parameters, and . If and are independent, then
has an distribution. It is nonnegative and right-skewed.
Key results
For a normal distribution, the empirical rule says that about 68% of values lie within one standard deviation of the mean, 95% within two, and 99.7% within three. Exact probabilities come from the standard normal cumulative distribution function .
The normal distribution approximates the binomial distribution when and are sufficiently large. If , then
Because the binomial is discrete and the normal is continuous, a continuity correction often improves approximation:
where is the approximating normal random variable.
The distribution approaches the standard normal as degrees of freedom increase. Small samples have more uncertainty in the estimated standard deviation, so the distribution has heavier tails. This produces wider confidence intervals and larger critical values.
Chi-square and distributions are built from squared standardized quantities. That is why they are nonnegative and skewed. In categorical tests, a chi-square statistic accumulates squared discrepancies between observed and expected counts:
In ANOVA, an statistic compares mean squares:
Large chi-square or statistics often indicate evidence against a null model, but "large" must be judged relative to the appropriate degrees of freedom.
These distributions also differ in how their tails are used. Normal and tests for means often use one or two tails depending on the alternative hypothesis. Chi-square goodness-of-fit and independence tests use the upper tail because larger squared discrepancies are more inconsistent with the null; a statistic near zero means observed counts are unusually close to expected counts, not evidence against the model in the usual test. ANOVA tests also use the upper tail because the statistic is a ratio of explained to unexplained variance. Remembering the statistic's construction is more reliable than memorizing tail directions.
Degrees of freedom can be read as the amount of independent information left after estimating constraints. In a one-sample test, estimating the sample mean uses one constraint, leaving degrees of freedom for variability. In a chi-square table, fixed totals reduce how many cells can vary freely. This interpretation helps explain why larger samples make distributions look more normal and why changing the number of categories changes the chi-square reference curve.
Visual
| Distribution | Support | Shape | Key parameters | Common statistic |
|---|---|---|---|---|
| Normal | all real numbers | symmetric bell | ||
| all real numbers | symmetric, heavy tails | degrees of freedom | ||
| Chi-square | right-skewed | degrees of freedom | ||
| right-skewed | numerator and denominator df | ratio of mean squares |
Relative shape sketch
normal and t: chi-square and F:
^ ^
| normal |\
| /\ | \
| t / \ | \___
| ___/ \___ | \____
+----------------> +---------------->
0 0
Worked example 1: Normal probability with standardization
Problem: Suppose adult commute times in a region are approximately normal with mean minutes and standard deviation minutes. What proportion of commuters take longer than 45 minutes? What commute time marks the 90th percentile?
Method:
- Standardize 45 minutes:
- Convert to an upper-tail probability:
- From a standard normal table or software, .
- For the 90th percentile, find such that . The value is about 1.2816.
- Transform back:
Answer: About 4.0% of commuters take longer than 45 minutes. The 90th percentile is about 41.3 minutes.
Checked answer: The 90th percentile should be above the mean, and 41.3 is about 1.28 standard deviations above 31, which matches the standard normal percentile.
Worked example 2: Choosing , chi-square, or
Problem: For each study, identify the appropriate reference distribution and degrees of freedom where possible.
- A sample of 16 batteries has mean lifetime 52.1 hours and sample standard deviation 4.8 hours. Test whether the population mean differs from 50 hours.
- A genetics activity compares observed counts of four phenotype categories with expected Mendelian proportions.
- A one-way ANOVA compares mean exam scores across three teaching methods with 10 students per method.
Method:
- The battery problem involves one sample mean and unknown population standard deviation. Use a one-sample distribution with
The statistic is
- The phenotype problem compares observed and expected categorical counts. Use a chi-square goodness-of-fit statistic. With four categories and no estimated parameters, the degrees of freedom are
- The ANOVA problem compares three group means. There are groups and total observations. Use an distribution with
Answer: The correct reference distributions are for the one-sample mean, for the four-category goodness-of-fit test, and for the one-way ANOVA.
Checked answer: The chosen distributions match the statistic types: standardized mean with estimated standard error, squared categorical discrepancies, and ratio of between-group to within-group mean squares.
Code
from scipy import stats
# Normal probability and percentile
mu, sigma = 31, 8
p_long = 1 - stats.norm.cdf(45, loc=mu, scale=sigma)
p90 = stats.norm.ppf(0.90, loc=mu, scale=sigma)
print(f"P(commute > 45) = {p_long:.4f}")
print(f"90th percentile = {p90:.2f}")
# t statistic and two-sided p-value for battery example
t_stat = (52.1 - 50) / (4.8 / (16 ** 0.5))
p_value = 2 * (1 - stats.t.cdf(abs(t_stat), df=15))
print(f"t = {t_stat:.3f}, p = {p_value:.4f}")
# Critical F value for alpha = 0.05, df1 = 2, df2 = 27
print("F critical:", stats.f.ppf(0.95, dfn=2, dfd=27))
The code uses cumulative distribution functions for probabilities and percent point functions for percentiles or critical values. This is safer than copying table entries by hand, especially for nonstandard degrees of freedom.
Common pitfalls
- Using a normal critical value for a small-sample mean when the population standard deviation is unknown.
- Forgetting that chi-square and statistics cannot be negative.
- Treating degrees of freedom as decoration rather than part of the distribution.
- Applying a normal approximation to a binomial problem when expected successes or failures are too small.
- Forgetting a continuity correction when a more accurate normal approximation to a discrete count is needed.
- Comparing a statistic to the wrong tail of the reference distribution.