BackComprehensive Study Notes: Hypothesis Testing, t-Tests, ANOVA, Correlation, Regression, and Chi-Square
Study Guide - Smart Notes
Tailored notes based on your materials, expanded with key definitions, examples, and context.
Hypothesis Testing
Overview of Hypothesis Testing
Hypothesis testing is a statistical method used to make inferences or draw conclusions about a population based on sample data. It involves formulating a null hypothesis (H0) and an alternative hypothesis (H1), then using sample statistics to determine whether to reject H0.
Steps of Hypothesis Testing:
State the hypotheses (null and alternative).
Set the significance level (α).
Choose the appropriate test statistic.
Calculate the test statistic from sample data.
Determine the p-value or critical value.
Draw a conclusion (reject or fail to reject H0).
Types of Errors:
Type I Error (α): Rejecting a true null hypothesis.
Type II Error (β): Failing to reject a false null hypothesis.
Statistical Power: Probability of correctly rejecting a false null hypothesis (1 - β).
t-Tests
One-Sample t-Test
Used to compare the mean of a single sample to a known value or population mean.
Test Statistic:
Interpretation: Compare calculated t to critical t-value from t-distribution table.
Effect Size: Cohen's d can be used to measure effect size.
Independent Samples t-Test
Compares the means of two independent groups to determine if they are significantly different.
Equal Sample Sizes:
Pooled Standard Deviation:
Different Sample Sizes:
Related (Paired) Samples t-Test
Used when the same subjects are measured twice (e.g., before and after treatment).
Test Statistic:
Where: is the mean of the difference scores, is the standard deviation of the difference scores.
Analysis of Variance (ANOVA)
One-Way ANOVA
Used to compare means across three or more independent groups.
F Statistic:
Mean Square Between:
Mean Square Within:
Interpretation: Large F indicates significant differences among group means.
Two-Way ANOVA
Used to examine the effect of two independent variables (factors) on a dependent variable, including possible interaction effects.
Main Effects: Effect of each factor independently.
Interaction Effect: Combined effect of both factors.
F Statistic for Each Effect:
Repeated Measures ANOVA
Used when the same subjects are measured under different conditions or at different times.
Accounts for within-subject variability.
Correlation
Pearson Correlation Coefficient (r)
Measures the strength and direction of the linear relationship between two continuous variables.
Formula:
df for correlation:
Interpretation: r ranges from -1 (perfect negative) to +1 (perfect positive).
Linear Regression
Simple Linear Regression
Models the relationship between a dependent variable (Y) and an independent variable (X) using a straight line.
Regression Equation:
Slope (b1):
Intercept (b0):
Interpretation: Slope indicates the change in Y for a one-unit increase in X.
Chi-Square Tests
Chi-Square Goodness of Fit Test
Tests whether observed frequencies differ from expected frequencies in one categorical variable.
Test Statistic:
O: Observed frequency
E: Expected frequency
Chi-Square Test for Independence
Tests whether two categorical variables are independent.
Same formula as above, applied to contingency tables.
Summary Table: Hypothesis Tests and Formulas
Test | Purpose | Formula |
|---|---|---|
One-Sample t-Test | Compare sample mean to population mean | |
Independent Samples t-Test | Compare means of two independent groups | |
Related Samples t-Test | Compare means of paired samples | |
One-Way ANOVA | Compare means of 3+ groups | |
Pearson Correlation | Measure linear relationship | |
Linear Regression | Predict Y from X | |
Chi-Square | Test categorical association |
Additional info:
Effect sizes (e.g., Cohen's d, eta squared) are important for interpreting the magnitude of results.
SPSS is a common statistical software used for performing these analyses.
Post-hoc tests are used after ANOVA to determine which groups differ.