scoresvideos
AP Statistics
One-page, printable cheatsheet
Cheatsheet visualization
Table of Contents

๐Ÿ“Šap statistics review

6.7 Potential Errors When Performing Tests

Verified for the 2025 AP Statistics examโ€ขCitation:

What is an Error?

No matter how well we design our test, perform our calculations and follow our correct procedures, we are still prone to error in our tests. This doesn't necessarily mean we did something wrong in our sampling or our calculations, but just that our calculations gave us an incorrect result. We always have that small random chance of achieving a rare sample that leads us to incorrect results and there are ways that we can minimize this effect. ๐Ÿ€

Source: AB Tasty

In inferential statistics, there are two types of errors that can occur in a hypothesis test: type I error and type II error.

Type I Error

type I error occurs when we reject the null hypothesis when it is actually true. This error is also known as a "false positive." This is due to a low p-value that lead us to make a decision, but actually we drew an extremely rare sample from our population. The probability of a type I error is represented by the significance level ฮฑ, which is the probability that we will reject the null hypothesis when it is true. In general, we set ฮฑ to be a small value, such as 0.01 or 0.05, in order to minimize the probability of a type I error. โž• 

Type II Error

type II error occurs when we fail to reject the null hypothesis when it is actually false. This error is also known as a "false negative." This is due to the fact that we did not get a low enough p-value to reject our Ho, but in reality, our Ho is not the truth and there should have been convincing evidence for Ha. The probability of a type II error is represented by ฮฒ, which is the probability that we will fail to reject the null hypothesis when it is false. โž–

Again, 0.05 is a good significance level that minimizes the probability of making this type of error, while also being sure that our calculations obtained are still statistically significant. The probability of making a Type 2 Error is ๐žซ. This is easy to remember because the probabilities of a Type 1/2 Error are alpha/beta respectively.

The probability of a Type II error decreases when any of the following occurs, provided the others do not change: 

  1. Sample size(s) increases.
  2. Significance level (ฮฑ) of a test increases.
  3. Standard error decreases.
  4. True parameter value is farther from the null.

Power

There are several ways that we can minimize the probability of both type I and type II errors. One way is to use a larger sample size, as this can increase the power of the test, which is the probability of correctly rejecting the null hypothesis when it is false. We can also choose a more stringent significance level, such as ฮฑ = 0.01, in order to decrease the probability of a type I error. However, this also increases the probability of a type II error (1 - power). ๐Ÿ‘Š

A significance level of 0.05 is usually a good middle ground that minimizes type II error, but also keeps us from making other errors in our study.

The complement of ๐žซ is known as the power of our test. The power is basically a way of saying how strong our test is because it is the probability of NOT making a Type 2 error. We can increase our power by increasing our sample size. Remember, the larger our sample, the closer our estimate is the population parameter, so the less likely we are to make a mistake.

Source: VWO

Test Pointers

Common AP Test questions regarding types of errors and power typically ask the following questions: ๐Ÿค”

Identify Error

The first thing AP is likely to ask is how to identify either a Type 1 or 2 error. This is basically writing out the definitions above in context of the given problem. Learn the definitions using the trick above and this part is easy.

Consequence of Error

Past AP Statistics tests also loved asking about the consequence(s) of an error. If we rejected a null hypothesis when we shouldn't have, what are the consequences (in context) of making such an error?

Increase Power

The last thing that AP likes to ask about regarding errors and power is how we can increase power. The answer is always to increase sample size.

Example

image courtesy of:  pixabay.com

In a recent study, a researcher was testing the claim that 85% of people are satisfied with their current personal reading goals and achievements. The researcher has reason to believe that this proportion is lower and that people are actually not happy with their personal reading plans and need a new library to borrow from. Therefore, the researcher tests the following hypotheses to see if opening a new public library would help people reach their personal reading goals:

Ho: p = 0.85

Ha: p < 0.85

a) Describe a Type 2 error in context of the problem. Also, list a consequence of making this type of error.

If the researcher makes a Type 2 error in this problem, he/she has failed to reject the Ho, when in fact it should be rejected. This means that the researcher concluded that we did not have evidence that the true population proportion was less than 0.85, when in fact, there is convincing evidence that it is less than 0.85. A consequence of this error is that people will likely remain largely unhappy with their reading achievement when a new library may help them reach their reading goals.

b) What can the researcher do to increase the power of this test?

The researcher can increase the power of this test and therefore decrease the probability of making a Type 2 error by increasing the sample size in the study.

๐ŸŽฅ Watch: AP Stats - Inference: Errors and Powers of Test 

Key Terms to Review (10)

Alternative Hypothesis: The alternative hypothesis is a statement that contradicts the null hypothesis, suggesting that there is an effect, a difference, or a relationship present in the data. This hypothesis is what researchers aim to support through their analysis and testing, as it represents the possibility that something interesting is happening beyond random chance.
Hypothesis Test: A hypothesis test is a statistical method used to make inferences or draw conclusions about a population based on sample data. This process involves formulating a null hypothesis and an alternative hypothesis, followed by analyzing the data to determine whether to reject the null hypothesis in favor of the alternative. The validity and reliability of these tests are crucial for assessing claims about population parameters and making data-driven decisions.
Null Hypothesis: The null hypothesis is a statement that assumes there is no effect or no difference in a given situation, serving as the foundation for statistical testing. It provides a baseline against which alternative hypotheses are tested, guiding researchers in determining whether observed data significantly deviates from what is expected under this assumption.
P-value: A P-value is a measure used in hypothesis testing to determine the strength of evidence against the null hypothesis. It quantifies the probability of observing test results at least as extreme as the ones obtained, assuming that the null hypothesis is true. A smaller P-value indicates stronger evidence against the null hypothesis, which is crucial for decision-making in various statistical tests.
Power of the Test: The power of a statistical test is the probability that it correctly rejects a false null hypothesis, thus detecting an effect when there is one. High power indicates a greater likelihood of identifying a true effect in the data, making it essential for assessing the effectiveness of a test. This concept connects deeply with the likelihood of making Type II errors and is influenced by factors such as sample size, effect size, and significance level.
Sample Size: Sample size refers to the number of observations or data points collected from a population for the purpose of statistical analysis. It plays a critical role in determining the reliability and validity of the results, impacting the precision of estimates and the power of hypothesis tests.
Significance Level: The significance level, often denoted as alpha (\(\alpha\)), is the threshold used to determine whether to reject the null hypothesis in statistical hypothesis testing. It represents the probability of making a Type I error, which occurs when a true null hypothesis is incorrectly rejected. Understanding the significance level is crucial for interpreting results and making informed decisions based on statistical tests.
Standard Error: Standard Error is a statistic that measures the accuracy with which a sample represents a population, specifically quantifying the variability of a sample mean from the population mean. It plays a critical role in constructing confidence intervals and conducting hypothesis tests, helping to assess how much sample means are expected to fluctuate around the true population mean. A smaller standard error indicates that the sample mean is a more precise estimate of the population mean.
Type II error: A Type II error occurs when a statistical test fails to reject a false null hypothesis, meaning that the test concludes there is no effect or difference when, in fact, one exists. This error is crucial in hypothesis testing, as it can lead to missed opportunities to identify significant effects or differences between populations.
Type I Error: A Type I Error occurs when a true null hypothesis is incorrectly rejected, leading to the conclusion that there is an effect or difference when, in fact, none exists. This error is also known as a false positive and is critical to understand in the context of hypothesis testing, as it reflects the risk of making a wrong decision based on sample data.