Skip to content

Normal Distribution

Approximation

Models the number of successes \(X\) in \(n\) independent Bernoulli trials, each with success probability \(p\).

\[ P(X = k) = \binom{n}{k} p^k (1 - p)^{n - k} \]
  • Given,
    • \(X \sim \text{Binomial}(n, p)\)
    • Mean: \(\mu = np\)
    • Variance: \(\sigma^2 = np(1 - p) = npq\)

Normal Approximation

For large \(n\), the binomial distribution approximates a normal distribution:

\[ X \approx \mathcal{N}(np, npq) \]
  • Bell curve shape
  • Approximation becomes accurate for \(n \gtrsim 30\) and moderate \(p\)
  • Bernoulli Trials:

    Single trial with two outcomes → success (probability \(p\)) or failure (probability \(1 - p\))

Equation

The probability density function (PDF) of a normal (Gaussian) distribution with mean \(\mu\) and standard deviation \(\sigma\) is:

\[ P(X = x) = \frac{1}{\sigma\sqrt{2\pi}} \exp\left( -\frac{(x - \mu)^2}{2\sigma^2} \right) \]

Where:

  • \(\mu = n \cdot p\) is the mean (expected value),
  • \(\sigma^2 = n \cdot p \cdot q\) is the variance,
  • \(\sigma = \sqrt{n \cdot p \cdot q}\) is the standard deviation,
  • \(q = 1 - p\) is the probability of failure.

This formula gives a continuous approximation of the binomial distribution when \(n\) is large and \(p\) is not too close to 0 or 1.

Descriptive Statistics

Type Formula
Support \(x \in (-\infty, \infty)\)
Mean \(\mathbb{E}[X] = \mu\)
Variance \({Var}(X) = \sigma^2\)
Standard Deviation \(\sigma\)
Probability Density Function (PDF) \(f(x) = \frac{1}{\sigma \sqrt{2\pi}} \exp\left(-\frac{(x-\mu)^2}{2\sigma^2}\right)\)
Cumulative Distribution Function (CDF) \(F(x) = \mathbb{P}(X \leq x) = \int_{-\infty}^x f(t) dt\)

Intuition and Use

  • Why Approximate Binomial with Normal?
    • Binomial becomes computationally intensive for large \(n\)
    • Normal is continuous and easy to compute probabilities for \((e.g., P(20 ≤ X ≤ 50))\). This would be a monster to compute using Binomial distribution.
    • Better for analysis and visualization
    • Normal distributions have well-known properties for sums, differences, and linear combinations. This makes complex probability problems much more manageable.
  • When Does Approximation Work Well?
    • \(n\) is large (rule of thumb: \(n \geq 30\) or so). The distributions agree well with a lot of overlap.
    • \(p\) not too close to 0 or 1 (i.e., \(np(1 - p)\) not small)
    • Fails when \(p\) or \(q\) is too small, may predict impossible values (like negative successes)
  • Central Limit Theorem (CLT):
    • Sum of many i.i.d. random variables (of any distribution with finite variance) converges to a normal distribution
    • Binomial is a sum of Bernoulli trials → classic CLT case

Probability Density Function (PDF) and Cumulative Distribution Function (CDF)

  • PDF describes the probability density at each point xx; area under the curve between two points gives the probability that \(X\) lies within that range.
  • CDF is the integral of the PDF from \(\int_{-\infty}^x f(x) dx\), representing the probability that \(X \leq x\).
  • The standard normal distribution (\(\mu=0, \sigma=1\)) is the canonical form, often denoted \(\Phi(x)\) for the CDF.

Calculating Probabilities

  • Probability \(P(a \leq X \leq b) = \int_a^b f(x) dx = \Phi(b) - \Phi(a)\), where \(f(x)\) is the PDF.
  • CDF values can be found using lookup tables historically, now easily computed with Python libraries like SciPy.

Applying Normal Approximation to Binomial

  • Example: Flip a fair coin 400 times \((n=400, p=0.5)\):
    • Mean \(\mu = np = 200\)
    • Variance \(\sigma^2 = np(1-p) = 100\)
    • Standard deviation \(\sigma = 10\)
  • Calculating let’s say getting # of heads between → \(P(190 \leq X \leq 230)\) binomially requires summing 51 probabilities, cumbersome to compute directly.
  • Using normal approximation, calculate \(\Phi(230) - \Phi(190)\) to get probability efficiently.