Introduction to Discrete Random Variables

Welcome to one of the most important chapters in Statistics 1! Up until now, you might have been looking at data that has already happened (like heights of students in a class). In this chapter, we shift our focus to models—mathematical ways to predict what might happen in the future.

Understanding Discrete Random Variables is like learning the rules of a game before you play it. It allows us to calculate averages and risks for events that haven't occurred yet, which is exactly how insurance companies, game designers, and scientists make decisions!

1. What is a Discrete Random Variable?

Let’s break this title down into two simple parts:

1. Discrete: This means the outcomes are distinct and separate. You can count them (like 1, 2, 3...). You can't have "half" an outcome in these cases. For example, the number of heads in 3 coin flips is discrete (you can't get 1.5 heads).
2. Random Variable: This is a quantity whose value depends on the outcome of a random event. We usually use a capital letter like \(X\) to represent the "name" of the variable and a lowercase \(x\) to represent the actual value it takes.

Analogy: Imagine a vending machine. The machine itself is the Random Variable \(X\). The specific snack that falls out (a chocolate bar, chips, or crackers) is the value \(x\). Because you don't know exactly which one will drop, it's "random."

Key Terms to Remember:

Sample Space: The list of all possible outcomes (e.g., for a die, it is {1, 2, 3, 4, 5, 6}).
Probability Distribution: A full description (usually a table) showing every possible outcome and its probability.

Quick Review: For any valid probability distribution, the sum of all probabilities must equal 1. \( \sum P(X = x) = 1 \).

2. The Probability Function and the CDF

The Probability Function \(p(x)\)

The probability function, written as \(P(X = x)\), tells us the chance of a specific outcome happening. Sometimes this is given as a formula, like \(P(X = x) = kx\). To find the value of \(k\), just remember the "Sum to 1" rule!

The Cumulative Distribution Function (CDF)

The CDF is written as \(F(x)\). Think of "cumulative" as a running total. It tells us the probability that the variable is less than or equal to a certain value.

Formula: \(F(x_0) = P(X \le x_0) = \sum_{x \le x_0} p(x)\)

Example: If you are rolling a die, \(F(2)\) is the probability of rolling a 1 OR a 2.
\(F(2) = P(X=1) + P(X=2) = \frac{1}{6} + \frac{1}{6} = \frac{2}{6}\).

Key Takeaway: \(P(X = x)\) is the probability for one specific point; \(F(x)\) is the accumulated probability up to that point.

3. Expected Value: The "Mean" of \(X\)

The Expected Value, written as \(E(X)\) or the Greek letter \(\mu\) (mu), is the long-term average value you would expect if you repeated the experiment many, many times.

How to calculate \(E(X)\):
1. Multiply each value \(x\) by its probability \(P(X=x)\).
2. Add all those results together.

Formula: \(E(X) = \sum x \cdot P(X = x)\)

Don't worry if this seems tricky at first! Just think of it as a "weighted average." If there is a 90% chance of winning $1 and a 10% chance of winning $100, the "Expected Value" is higher than $1 because the $100 "pulls" the average up, even though it happens rarely.

Did you know? The Expected Value doesn't have to be a value the variable can actually take. For a fair die, the \(E(X)\) is 3.5, even though you can never actually roll a 3.5!

4. Variance and Standard Deviation

While the Mean tells us the center, the Variance, written as \(Var(X)\) or \(\sigma^2\), tells us how spread out the values are.

To find the Variance, we use this very common formula:
\(Var(X) = E(X^2) - [E(X)]^2\)

Step-by-Step for Variance:

1. Find \(E(X)\) (the mean). Then square it.
2. Find \(E(X^2)\): Square each \(x\) value first, then multiply by its probability, and sum them up.
3. Subtract: \(E(X^2)\) minus the square of the mean.

Memory Trick: Think "The Mean of the Squares minus the Square of the Mean."

Common Mistake: Many students forget to square the mean at the end. Always double-check your subtraction step!

Standard Deviation: This is simply the square root of the Variance. \(\sigma = \sqrt{Var(X)}\).

5. Coding: Linear Transformations

Sometimes we change our data by adding a constant or multiplying by a factor (e.g., converting a score into a percentage). We call this coding.

If we have a new variable \(Y = aX + b\), here is what happens:

1. The Mean is affected by everything:
\(E(aX + b) = aE(X) + b\)
(If you double everyone's score and add 5, the average doubles and increases by 5.)

2. The Variance is ONLY affected by the multiplier (and it gets squared!):
\(Var(aX + b) = a^2 Var(X)\)
(Adding 5 to every score doesn't change how "spread out" they are, so the '+ b' disappears. The multiplier 'a' is squared because variance is a squared measure.)

Quick Summary:
• Mean: Follows the rule exactly.
• Variance: Ignore the addition/subtraction, square the multiplier.

6. The Discrete Uniform Distribution

This is a special, "fair" distribution. A Discrete Uniform Distribution occurs when every possible outcome has the exact same probability.

Example: Rolling a fair 6-sided die. Every number from 1 to 6 has a probability of \(\frac{1}{6}\).
If a random variable \(X\) can take values \(1, 2, ..., n\), then:
• \(P(X = x) = \frac{1}{n}\) for all \(x\).
• The Mean \(E(X)\) will be exactly in the middle: \(\frac{n+1}{2}\).

Key Takeaway: When you see the word "Uniform," think "Equal" or "Fair." This simplifies your calculations because you don't need a complex table—you know every probability is the same!

Summary Checklist:
• Do all my probabilities add up to 1?
• Did I remember to square the mean when calculating Variance?
• When coding \(Var(aX+b)\), did I remember to square the \(a\) and ignore the \(b\)?
• Is my \(E(X)\) somewhere between my smallest and largest \(x\) values? (If not, check your math!)