When comparing two groups, we often want to know if their averages differ. This method helps us do that when we don't know how spread out the data is in each group. It uses a special formula to calculate a "t-score" that tells us how different the groups are.
The t-score takes into account the averages, variability, and sizes of both groups. We then use this score to figure out if the difference is real or just by chance. This helps us make decisions about things like which teaching method works better or if a new drug is effective.
Comparing Two Population Means with Unknown Standard Deviations
T-score calculation for population means
- Compares means of two populations with unknown standard deviations
- Assumes populations are normally distributed (bell-shaped curve)
- Assumes samples are independent (not related or influencing each other)
- T-score test statistic formula: $t = \frac{\bar{x}_1 - \bar{x}_2}{\sqrt{\frac{s_1^2}{n_1} + \frac{s_2^2}{n_2}}}$
- $\bar{x}_1$ and $\bar{x}_2$ represent sample means (averages of each group)
- $s_1^2$ and $s_2^2$ represent sample variances (measure of variability within each group)
- $n_1$ and $n_2$ represent sample sizes (number of observations in each group)
- Null hypothesis ($H_0$): population means are equal ($\mu_1 = \mu_2$)
- Alternative hypothesis ($H_a$) options:
- Two-tailed: means are not equal ($\mu_1 \neq \mu_2$)
- Left-tailed: mean of population 1 is less than mean of population 2 ($\mu_1 < \mu_2$)
- Right-tailed: mean of population 1 is greater than mean of population 2 ($\mu_1 > \mu_2$)
- P-value: probability of obtaining test results at least as extreme as observed, assuming null hypothesis is true
Degrees of freedom in t-distributions
- Degrees of freedom (df) for comparing two population means calculated using Welch-Satterthwaite equation
- Formula: $df = \frac{(\frac{s_1^2}{n_1} + \frac{s_2^2}{n_2})^2}{\frac{(s_1^2/n_1)^2}{n_1-1} + \frac{(s_2^2/n_2)^2}{n_2-1}}$
- Calculated df rounded down to nearest integer (whole number)
- df determines shape of t-distribution
- Smaller df results in heavier tails compared to normal distribution (more spread out)
- As df increases, t-distribution approaches normal distribution (bell-shaped curve)
- Examples:
- Comparing test scores of two classes (Class A and Class B)
- Analyzing effectiveness of two different teaching methods (Traditional vs. Innovative)
Cohen's d for effect size
- Measures magnitude of difference between two population means
- Quantifies practical significance of difference
- Cohen's d formula: $d = \frac{\bar{x}_1 - \bar{x}_2}{s_p}$
- $\bar{x}_1$ and $\bar{x}_2$ represent sample means
- $s_p$ is pooled standard deviation, calculated as: $s_p = \sqrt{\frac{(n_1-1)s_1^2 + (n_2-1)s_2^2}{n_1 + n_2 - 2}}$
- Interpretation guidelines:
- Small effect size: $d = 0.2$
- Medium effect size: $d = 0.5$
- Large effect size: $d = 0.8$
- Independent of sample size, allows comparison across studies
- Examples:
- Comparing effectiveness of two drugs (Drug A vs. Drug B) on reducing blood pressure
- Evaluating impact of two exercise programs (Cardio vs. Strength) on weight loss
Statistical Inference and Error Types
- Confidence interval: range of values likely to contain the true population parameter
- Type I error: rejecting the null hypothesis when it is actually true (false positive)
- Type II error: failing to reject the null hypothesis when it is actually false (false negative)
- Statistical power: probability of correctly rejecting a false null hypothesis