Welcome to the Normal Distribution!
Ever noticed how most people are of average height, while very few are extremely tall or extremely short? Or how most apples in a bag weigh about the same, with only a few outliers? This natural pattern is called the Normal Distribution, often nicknamed the "Bell Curve."
In this chapter, we will learn how to model these real-world patterns using mathematics. Don’t worry if statistics feels a bit abstract—once you see the visual patterns, it all starts to click!
1. Continuous Random Variables
Before we dive into the curve, we need to understand what we are measuring. Unlike the Binomial Distribution (which counts "how many"), the Normal Distribution deals with Continuous Random Variables.
- Discrete (Binomial): Countable things (e.g., number of students, number of heads in a coin toss).
- Continuous (Normal): Measurable things that can take any value in a range (e.g., your exact height, the time taken to run a race, the weight of a chocolate bar).
Key Takeaway: If you measure it with a ruler, a scale, or a stopwatch, it’s likely a continuous variable!
2. The Normal Distribution: The "Bell Curve"
A Normal Distribution is defined by two important numbers (parameters):
- Mean (\(\mu\)): The center of the curve (the average).
- Variance (\(\sigma^2\)): How spread out the data is. (Note: \(\sigma\) is the Standard Deviation).
We write this as: \(X \sim N(\mu, \sigma^2)\)
Properties of the Bell Curve:
- Symmetry: The left side is a mirror image of the right side. The peak is exactly at the mean (\(\mu\)).
- Total Area = 1: Because the total probability of all outcomes must equal 1 (or 100%).
- Asymptotic: The "tails" of the curve get closer and closer to the horizontal axis but never actually touch it.
Analogy: Think of the curve like a pile of sand. Most of the sand is in the big heap in the middle (the mean), and it thins out as you move away to the left or right.
Quick Review:
If \(X \sim N(50, 25)\), then the mean is 50 and the variance is 25. This means the standard deviation (\(\sigma\)) is \(\sqrt{25} = 5\).
3. The Standard Normal Distribution (Z)
Since there are infinite possible values for \(\mu\) and \(\sigma\), mathematicians created a "gold standard" curve to make calculations easier. This is the Standard Normal Distribution, denoted by the letter \(Z\).
For the \(Z\)-distribution:
\(Z \sim N(0, 1)\) (Mean = 0, Variance = 1)
Standardization Formula:
To turn any "Normal" value (\(x\)) into a "Standard" value (\(z\)), we use:
\(Z = \frac{X - \mu}{\sigma}\)
This tells you how many standard deviations a value is away from the mean.
- A positive Z means the value is above average.
- A negative Z means the value is below average.
4. Finding Probabilities and Values
In H1 Math, you will primarily use your Graphing Calculator (GC) to find probabilities. You don't need to do complex integration!
A. Finding Probability \(P(X < x_1)\)
Use the normCdf function on your GC. You will need to input the Lower Bound, Upper Bound, \(\mu\), and \(\sigma\).
Common Mistake: Students often forget that the GC asks for \(\sigma\) (Standard Deviation), but the question might give you \(\sigma^2\) (Variance). Always square root the variance before typing it in!
B. Finding the Value \(x_1\) given Probability
If the question gives you the area (probability) and asks for the "cut-off" value, use invNorm on your GC.
The Beauty of Symmetry:
Because the curve is perfectly symmetrical:
1. \(P(X > \mu) = 0.5\)
2. \(P(X < \mu - a) = P(X > \mu + a)\)
3. \(P(X < x) = 1 - P(X > x)\)
Did you know? About 68% of all data in a normal distribution lies within 1 standard deviation of the mean. About 95% lies within 2 standard deviations!
5. Linear Combinations of Random Variables
Sometimes, we need to add or subtract variables. For example, if a cup of coffee (\(X\)) and a saucer (\(Y\)) are both normally distributed, what is the distribution of their total weight (\(X+Y\))?
A. Single Variable Scaling: \(aX + b\)
If \(X \sim N(\mu, \sigma^2)\), then for the new variable \(W = aX + b\):
- New Mean: \(E(aX + b) = aE(X) + b\)
- New Variance: \(Var(aX + b) = a^2Var(X)\)
Tip: Notice how the "+ b" affects the mean but disappears in the variance? Shifting the whole curve left or right doesn't change how "spread out" it is! Also, always square the "a" when calculating variance.
B. Combining Two Independent Variables: \(aX + bY\)
If \(X\) and \(Y\) are independent:
- New Mean: \(E(aX + bY) = aE(X) + bE(Y)\)
- New Variance: \(Var(aX + bY) = a^2Var(X) + b^2Var(Y)\)
CRITICAL RULE: When subtracting variables (e.g., \(X - Y\)), you STILL ADD the variances!
\(Var(X - Y) = Var(X) + Var(Y)\)
Why? Because combining two uncertain things always creates MORE uncertainty (spread), never less.
Key Takeaway:
Mean follows the sign (add if adding, subtract if subtracting). Variance always adds and always squares the coefficients.
6. Summary and Tips for Success
- Sketch the curve: Always draw a quick bell curve and shade the area you are looking for. It prevents silly mistakes!
- Check your parameters: Read carefully—did the question give you \(\sigma\) or \(\sigma^2\)?
- Z-score is your friend: If the mean or variance is unknown, you must standardize to \(Z\) to solve the problem.
- Independence: You can only add variances if the variables are independent. Look for this keyword in the question.
Don't worry if this seems tricky at first! The Normal Distribution is one of the most consistent topics. Once you master the GC steps and the symmetry rules, you'll be able to tackle almost any question.