Normal distribution

Log in to access the full course.

The bell curve

The normal distribution (also called the Gaussian distribution) is the most important distribution in statistics. It describes many naturally occurring measurements — heights, test scores, measurement errors — and plays a central role in statistical theory.

Its shape is a symmetric, bell-shaped curve centered at its mean μ\mu, with spread controlled by its standard deviation σ\sigma.

The parameters

A normal distribution is fully described by two parameters:

  • μ\mu (mu): the mean — the center of the distribution.
  • σ\sigma (sigma): the standard deviation — how spread out the values are. Larger σ\sigma means a wider, flatter bell.

We write XN(μ,σ2)X \sim \mathcal{N}(\mu, \sigma^2) to say "XX follows a normal distribution with mean μ\mu and variance σ2\sigma^2."

The probability density function is:

f(x)=1σ2πexp((xμ)22σ2)f(x) = \frac{1}{\sigma\sqrt{2\pi}} \exp\left(-\frac{(x-\mu)^2}{2\sigma^2}\right)

You do not need to memorize this — but notice that it is symmetric around μ\mu and falls off as xx moves away from the mean.

The 68-95-99.7 rule

A practical rule for any normal distribution:

  • ~68% of values fall within 1 standard deviation of the mean (μ±σ\mu \pm \sigma)
  • ~95% fall within 2 standard deviations (μ±2σ\mu \pm 2\sigma)
  • ~99.7% fall within 3 standard deviations (μ±3σ\mu \pm 3\sigma)

Example: adult male heights in the US are approximately N(177 cm,72)\mathcal{N}(177\text{ cm},\, 7^2). So about 95% of men are between 17714=163177 - 14 = 163 cm and 177+14=191177 + 14 = 191 cm.

The standard normal

The standard normal distribution is the special case N(0,1)\mathcal{N}(0, 1) — mean 0, standard deviation 1. Any normal distribution can be converted to standard normal via standardization:

Z=XμσZ = \frac{X - \mu}{\sigma}

The value ZZ (called a z-score) tells you how many standard deviations XX is above or below the mean. A z-score of 22 means the value is 2 standard deviations above average.

Why the normal distribution is everywhere

The central limit theorem (CLT) explains its ubiquity: the sum (or average) of many independent random variables — regardless of their individual distributions — tends toward a normal distribution as the number of variables grows. Since many real-world quantities are the result of many small independent influences adding together, normality arises naturally.

Key properties

  • Symmetric around μ\mu: mean = median = mode.
  • Fully described by just μ\mu and σ\sigma.
  • The sum of independent normal random variables is also normal.
  • Tails extend to ±\pm\infty but become negligibly small beyond 3σ3\sigma.