Comparing Means: t-Tests
The t-test is one of the most commonly used statistical tests. It compares means to determine if there is a statistically significant difference. There are three types of t-tests, each for different research situations.
Types of t-Tests
One-Sample t-Test
Purpose: Compare a sample mean to a known or hypothesized population value
Research Question: Is the average IQ of students in our program different from 100?
H₀: μ = 100 (population mean equals 100)
H₁: μ ≠ 100 (population mean differs from 100)
When to Use:
- One group measured once
- Comparing to a known standard or benchmark
- Comparing to a theoretical value
Independent Samples t-Test
Purpose: Compare means between two different (unrelated) groups
Research Question: Do males and females differ in math anxiety?
H₀: μ₁ = μ₂ (no difference between groups)
H₁: μ₁ ≠ μ₂ (groups differ)
When to Use:
- Two separate groups
- Different people in each group
- Examples: treatment vs. control, male vs. female, experimental vs. comparison
Paired Samples t-Test
Purpose: Compare means from the same group at two different times or under two conditions
Research Question: Did anxiety decrease after the intervention?
H₀: μ_diff = 0 (no change from pre to post)
H₁: μ_diff ≠ 0 (significant change)
When to Use:
- Same participants measured twice
- Pre-test/post-test designs
- Matched pairs (twins, matched controls)
Assumptions of t-Tests
Continuous DV
The dependent variable must be measured at interval or ratio level
Check: Nature of your variable
Independence
Observations are independent (one person's score doesn't affect another's)
Check: Study design—how were data collected?
Normality
Data should be approximately normally distributed
Check: Histogram, Q-Q plot, Shapiro-Wilk test
Less important with large samples (n > 30)
Homogeneity of Variance
Groups should have similar variances (for independent t-test)
Check: Levene's test
If violated, use Welch's t-test
Interpreting t-Test Output
Sample SPSS Output (Independent Samples t-Test)
| Group | N | Mean | SD |
|---|---|---|---|
| Treatment | 45 | 78.4 | 12.3 |
| Control | 42 | 71.2 | 11.8 |
| t | df | p (2-tailed) | Mean Difference | 95% CI |
|---|---|---|---|---|
| 2.78 | 85 | .007 | 7.2 | [2.1, 12.3] |
How to Interpret:
- t-value (2.78): The test statistic; larger = bigger difference relative to variability
- df (85): Degrees of freedom (n₁ + n₂ - 2 for independent t-test)
- p-value (.007): Probability of this result if H₀ were true; p < .05 means significant
- Mean difference (7.2): Treatment group scored 7.2 points higher
- 95% CI [2.1, 12.3]: We're 95% confident the true difference is between 2.1 and 12.3
Effect Size: Cohen's d
Cohen's d measures the magnitude of the difference in standard deviation units. It tells you HOW BIG the effect is, not just whether it's significant.
d = (M₁ - M₂) / SD_pooled
Interpreting Cohen's d:
| d = 0.2 | Small effect | Noticeable but not dramatic |
| d = 0.5 | Medium effect | Moderate, meaningful difference |
| d = 0.8 | Large effect | Substantial difference |
Example Calculation:
M₁ = 78.4, M₂ = 71.2, SD_pooled ≈ 12.0
d = (78.4 - 71.2) / 12.0 = 7.2 / 12.0 = 0.60
Interpretation: Medium effect size
Reporting t-Test Results
APA Format:
"An independent-samples t-test was conducted to compare test scores between treatment and control groups. There was a significant difference in scores for the treatment group (M = 78.4, SD = 12.3) and the control group (M = 71.2, SD = 11.8); t(85) = 2.78, p = .007, d = 0.60. The treatment group scored significantly higher than the control group, with a medium effect size."
Always Include:
- ✓ Type of t-test used
- ✓ Means and SDs for each group
- ✓ t-value and degrees of freedom: t(df)
- ✓ Exact p-value (or < .001 if very small)
- ✓ Effect size (Cohen's d)
- ✓ Direction of the difference
Common Mistakes with t-Tests
- Using independent t-test for paired data: If same people measured twice, use paired t-test
- Running multiple t-tests: Comparing 3+ groups? Use ANOVA instead to avoid inflated Type I error
- Ignoring assumptions: Check normality and equal variances
- Forgetting effect size: Significance doesn't tell you how large the effect is