Skip to main content
Back

Sampling Distributions: Concepts, Properties, and Applications

Study Guide - Smart Notes

Tailored notes based on your materials, expanded with key definitions, examples, and context.

Chapter 6: Sampling Distributions

Introduction

This chapter introduces the concept of sampling distributions, a foundational topic in inferential statistics. Understanding sampling distributions is essential for making valid inferences about population parameters based on sample statistics.

Population Parameters and Sample Statistics

Definitions

  • Parameter: A numerical descriptive measure of a population. It is calculated using all observations in the population, and its value is almost always unknown.

  • Sample Statistic: A numerical descriptive measure of a sample. It is calculated using the observations in the sample.

Table: Population Parameters and Corresponding Sample Statistics

Population Parameter

Sample Statistic

Mean: μ

Sample mean:

Median: η

Sample median: M

Variance:

Sample variance:

Standard deviation:

Sample standard deviation: s

Proportion: p

Sample proportion:

The Concept of a Sampling Distribution

Definition

  • Sampling Distribution: The probability distribution of a sample statistic calculated from a sample of n measurements.

Sampling distributions describe the variability of a statistic from sample to sample and are central to statistical inference.

Comparing Estimators

  • Sample mean () and sample median (M) are both estimators of the population mean (μ).

  • Depending on the sample, either or M may be closer to μ.

Example: In one sample, may be closer to μ than M, while in another sample, M may be closer.

Sampling Distribution Example

  • For a sample of n = 25 temperature measurements, the sampling distribution of is approximately normal, centered at the population mean (μ).

Estimating Population Variance

  • Different statistics can be used to estimate the population variance (), each with its own sampling distribution.

Figure: Sampling distributions for two statistics (A and B) estimating may differ in spread and center.

Table: Outcomes for n = 2 Coin Tosses

Outcome (Toss 1, Toss 2)

Probability

H (x = 1), H (x = 1)

1/4

H (x = 1), T (x = 0)

1/4

T (x = 0), H (x = 1)

1/4

T (x = 0), T (x = 0)

1/4

Properties of Sampling Distributions

Unbiasedness and Minimum Variance

  • Estimator: A rule or formula that tells us how to use sample data to calculate a single number to estimate a population parameter.

  • Unbiased Estimator: An estimator whose sampling distribution has a mean equal to the parameter it estimates.

  • Biased Estimator: An estimator whose sampling distribution mean does not equal the parameter.

  • Minimum Variance: Among unbiased estimators, the one with the smallest variance is preferred.

Example: The sample mean is an unbiased estimator of the population mean μ.

The Sampling Distribution of and the Central Limit Theorem

Properties of the Sampling Distribution of

  • The mean of the sampling distribution of equals the mean of the sampled population:

  • The standard deviation of the sampling distribution of (standard error) is:

Central Limit Theorem (CLT)

  • If a random sample of n observations is selected from any population with mean μ and standard deviation σ, then for sufficiently large n, the sampling distribution of will be approximately normal with mean μ and standard deviation .

  • The larger the sample size, the better the normal approximation.

Example: For n = 25, the sampling distribution of is nearly normal, even if the population distribution is not normal.

Summary

  • Population parameters are typically unknown and estimated using sample statistics.

  • Sampling distributions describe the variability of sample statistics and are crucial for statistical inference.

  • Unbiasedness and minimum variance are desirable properties for estimators.

  • The Central Limit Theorem allows us to use normal probability models for sample means when sample sizes are large.

Pearson Logo

Study Prep