BackLecture 19: Binomial Distribution – Study Notes
Study Guide - Smart Notes
Tailored notes based on your materials, expanded with key definitions, examples, and context.
Overview of Discrete Probability Distributions
Introduction
This lecture focuses on the Bernoulli and Binomial distributions, which are fundamental discrete probability distributions in statistics. These distributions model experiments with two possible outcomes and are widely used in data analysis, quality control, and inferential statistics.
Review: Random Variables and Probability Distributions
Key Concepts
Random Variables (RV): Variables whose values are determined by the outcome of a random experiment.
Probability Distributions: Functions that assign probabilities to each possible value of a random variable.
Probability Mass Function (PMF): For discrete random variables, the PMF gives the probability that the variable takes a specific value.
Discrete vs. Continuous Random Variables: Discrete variables take countable values; continuous variables take values in an interval.
Mean and Variance: The mean (expected value) and variance measure the central tendency and spread of a random variable, respectively.
Rules for Variances
The variance of the sum of two random variables is not always the sum of their variances, especially if the variables are dependent.
Example: If X is the number of heads and Y the number of tails in 4 tosses of a fair coin, then X + Y = 4 always, so Var(X + Y) = 0 even though Var(X) = Var(Y) = 1.
General Rule: in general.
Bernoulli Distribution
Definition and Properties
The Bernoulli distribution models a single trial experiment with two possible outcomes: success (S) and failure (F).
Examples: Pass/Fail, Defective/Non-defective, Female/Male, Yes/No, Head/Tail.
Bernoulli Experiment: An experiment with a single trial and two mutually exclusive outcomes.
Parameter: , the probability of success.
Notation:
Probability Mass Function (PMF)
Let be the number of successes (0 or 1):
X | 0 | 1 |
|---|---|---|
f(x) = P(X = x) | 1 - p | p |
General PMF:
Mean and Variance of Bernoulli Distribution
Mean (Expected Value):
Variance:
Standard Deviation:
Binomial Distribution
Definition and Properties
The Binomial distribution models the number of successes in a fixed number of independent Bernoulli trials, each with the same probability of success.
Discrete probability distribution for
Describes: Repeated Bernoulli trials (sequence of successes and failures)
"Success" is a label for one of the two outcomes, not necessarily a positive result.
Conditions for a Binomial Experiment
The experiment is performed a fixed number of times ( trials).
The trials are independent.
Each trial has two mutually exclusive outcomes (success or failure).
The probability of success () is constant for each trial.
Notation
= number of trials
= probability of success on each trial
= probability of failure
= number of successes in trials ()
Notation:
Probability Mass Function (PMF) of Binomial Distribution
The probability of obtaining successes in independent trials:
where is the binomial coefficient.
Examples
Example 1: In a random sample of 20 car-owning households, with (probability of having three or more cars):
Probability exactly 5 have three or more cars:
Probability less than 4 have three or more cars:
Example 2: Drawing 4 pets with 40% cats and 60% dogs, with replacement. = number of cats in 4 draws.
Binomial setting: , ,
PMF: for
Table of probabilities:
X
0
1
2
3
4
f(x) = P(X = x)
0.13
0.35
0.35
0.15
0.03
Mean, Variance, and Standard Deviation of Binomial Distribution
Mean (Expected Value):
Variance:
Standard Deviation:
Reason: The binomial random variable is the sum of independent Bernoulli random variables.
Shape of the Binomial Distribution
The shape depends on the value of :
Right-skewed:
Symmetric:
Left-skewed:
As the number of trials increases, the binomial distribution becomes more bell-shaped (approximately normal).
Rule of Thumb: If , the distribution is approximately bell-shaped.
Binomial Probability Histograms
Histograms illustrate the probability distribution for different values of and .
For small or extreme , the distribution is skewed; for large $n$ and $p$ near 0.5, it is symmetric and bell-shaped.
The Empirical Rule and Binomial Random Variables
Empirical Rule
In a bell-shaped distribution, about 95% of all observations lie within two standard deviations of the mean.
Interval: to
Observations outside this interval are considered unusual (occur less than 5% of the time).
Example: Using the Empirical Rule
Suppose 35% of car-owning households have three or more cars. In a sample of 400, 162 have three or more cars. Is this unusual?
Compute expected value and standard deviation:
Interval: to
Since 162 is outside [120.9, 159.1], the result is considered unusual.
Summary Table: Bernoulli vs. Binomial Distribution
Property | Bernoulli | Binomial |
|---|---|---|
Number of Trials | 1 | n (fixed) |
Possible Outcomes | 2 (Success/Failure) | 0, 1, ..., n (number of successes) |
PMF | ||
Mean | ||
Variance |
Key Takeaways
The Bernoulli and Binomial distributions are foundational for modeling binary outcomes and repeated trials.
Understanding their properties, formulas, and applications is essential for statistical inference and hypothesis testing.
The shape of the binomial distribution approaches normality as the number of trials increases, especially when .