Open oldoc63 opened 1 year ago
The international baccalaureate degree programme (IBDP) is a worl-wide educational program. Each year, students in participating schools can take a standardized assessment to earn their degree. Total scores on this assessment range from 1-45. Based on data provided here, the average total score for all students who took this exam in May 2020 was around 29.92. The distribution of the scores looks something like this:
Imagine a hypothetical online school, Statistics Academy, that offers a 5-week test-preparation program. Suppose that 100 students who took the IBDP assessment in May 2020 were randomly chosen to participate in the first group of this program and that these 100 students earned an average score of 31.16 points on the exam -about 1.24 points higher than the international average. Are these students really outperforming their peers? Or could this difference be attributed to random chance?
Before attempting to answer this question, it is useful to reframe it in a way that is testable. Right now, our question ("Are the statistics academy students really outperforming their peers?") is not clearly defined. Objectively, this group of 100 students performed better than the general population -but answering "yes, they outperformed their peers" doesn't feel particularly satisfying.
The reason it's not satisfying is this: if we randomly choose ANY group of 100 students from the population of all test takers and calculate the average score for that sample, there's a 50% chance it will be higher than the population average. Observing a higher average for a single sample is not surprising.
Of course, large differences from the population average are less probable - if all the statistics academy students earned 45 points on the exam (the highest possible score), we'd probably be convinced that these students had a real advantage. The trick is to quantify when differences are "large enough" to convince us that these students are systematically different from the general population. We can do this by reframing our question to focus on population(s), rather than our specific sample(s).
A hypothesis test begins with two competing hypothesis about the population that a particular sample comes from -in this case, the 100 statistics academy students:
There is one more clarification we need to make in order to fully specify the alternative hypothesis. Notice that, so far, we have not said anything about the average score for "population 1 in the diagram above, other than it is NOT 29.92. We actually have three choices for what the alternative hypothesis says about this population average:
Now that we have our null hypothesis, we can generate a null distribution: the distribution (across repeated samples) of the statistic we are interested in if the null hypothesis is true. In the IBDP example described above, this is the distribution of average scores for repeated samples of size 100 drawn from a population with an average score of 29.92.
This might feel familiar if you've learned about the central limit theorem (CLT). Statistical theory allows us to estimate the shape of this distribution using a single sample. Let's simulate the null distribution using our population.
This yields the following distribution, which is approximately normal and centered at the population average:
If the null hypothesis is true, then the average score of 31.16 observed among statistics academy students is simply one of the values from this distribution. Let's plot the sample average on the null distribution. Notice that this value is within the range of the null distribution, but it is off to the right where there is less density:
Here is the basic question asked in our hypothesis test:
Given that null hypothesis is true (that the 100 students from statistics academy were sampled from a population with an average IBDP score of 29.92), how likely is it that their average score is 31.6?
The minor problem with this question is that the probability of any exact average score is very small, so we really want to estimate the probability of a range of scores. Let's now return to our three possible alternative hypotheses and see how the question and calculation each change slightly, depending on which one we choose:
Alternative Hypothesis: The sample of 100 scores earned by statistics academy students came from a population with an average score that is greater than 29.92.
In this case , we want to know the probability of observing a sample average greater than or equal to 31.16 given that the null hypothesis is true. Visually, this is the proportion of the null distribution that is greater than or equal to 31.16 (shaded in red below). Here, the red region makes up about 3.1% of the total distribution. This proportion, which is usually written in decimal form (i.e., 0.031), is called a p-value.
Alternative Hypothesis: The sample of 100 scores earned by statistics academy students came from a population with an average score that is not equal to (i.e., greater than or less than) 29.92.
We observed a sample average of 31.16 among the statistics academy students, which is 1.24 points higher than the population average (if the null hypothesis is true) of 29.92. In the first version of the hypothesis test (option 1), we estimated the probability of observing a sample average that is at least 1.24 points higher than the population average. For the alternative hypothesis described in option 2, we're interested in the probability of observing a sample average that is at least 1.24 points different (higher or lower) than the population average. Visually, this is the proportion of the null distribution that is at least 1.24 units away from the population average (shaded in red below). Note that this area is twice as large as in the previous example, leading to a p-value that is also twice as large: 0.062.
While option 1 is often referred to as One Sided or One-Tailed test, this version is referred to as Two Sided or Two-Tailed test, referencing the two tails of the null distribution that are counted in the p-value. It is important to note that most hypothesis testing functions in Python and R implement a two-tailed test by default.
Alternative Hypothesis: The sample of 100 scores earned by statistics academy students came from a population with an average score that is less than 29.92.
Here, we want to know the probability of observing a sample average less than or equal to 31.16, given that the null hypothesis is true. This is also a one-sided test, just the opposite side from option 1. Visually, this is the proportion of the null distribution that is less than or equal to 31.16. This is a very large area (almost 97% of the distribution!), leading to a p-value of 0.969.
At this point, you may be thinking: why would anyone ever choose this version of the alternative hypothesis?! In real life, if a test-prep program was planning to run a rigorous experiment to see whether their students are scoring differently than the general population, they should choose the alternative hypothesis before collecting any data. At that point, they won’t know whether their sample of students will have an average score that is higher or lower than the population average — although they probably hope that it is higher.
In the three examples above, we calculated three different p-values (0.031, 0.062, and 0.969, respectively). Consider the first p-value of 0.031. The interpretation of this number is as follows:
If the 100 students at Statistics Academy were randomly selected from the full population (which had an average score of 29.92), there is a 3.1% chance of their average score being 31.16 points or higher.
This means that it is relatively unlikely, but not impossible, that the Statistics Academy students scored higher (on average) than their peers by random chance, despite no real difference at the population level. In other words, the observed data is unlikely if the null hypothesis is true. Note that we have directly tested the null hypothesis, but not the alternative hypothesis! We therefore need to be a little careful about how we interpret this test: we cannot say that we’ve proved that the alternative hypothesis is the truth — only that the data we collected would be unlikely under null hypothesis, and therefore we believe that the alternative hypothesis is more consistent with our observations.
While it is perfectly reasonable to report a p-value, many data scientists use a predetermined threshold to decide whether a particular p-value is significant or not. P-values below the chosen threshold are declared significant and lead the data scientist to “reject the null hypothesis in favor of the alternative”. A common choice for this threshold, which is also sometimes referred to as Alpha, is 0.05 — but this is an arbitrary choice! Using a lower threshold means you are less likely to find significant results, but also less likely to mistakenly report a significant result when there is none.
Using the first p-value of 0.031 and a significance threshold of 0.05, Statistics Academy could reject the null hypothesis and conclude that the 100 students who participated in their program scored significantly higher on the test than the general population.
Note that different alternative hypotheses can lead to different conclusions. For example, the one-sided test described above (p = 0.031) would lead a data scientist to reject the null at a 0.05 significance level. Meanwhile, a two-sided test on the same data leads to a p-value of 0.062, which is greater than the 0.05 threshold. Thus, for the two-sided test, a data scientist could not reject the null hypothesis. This highlights the importance of choosing an alternative hypothesis ahead of time!
As you have hopefully learned from this article, hypothesis testing can be a useful framework for asking and answering questions about data. Before you use hypothesis tests in practice, it is important to remember the following:
A p-value is a probability, usually reported as a decimal between zero and one.
A small p-value means that an observed sample statistic (or something more extreme) would be unlikely to occur if the null hypothesis is true.
A significance threshold can be used to translate a p-value into a “significant” or “non-significant” result.
In practice, the alternative hypothesis and significance threshold should be chosen prior to data collection.
There are many different hypothesis tests that can be used to answer different kinds of questions about data. Now that you’ve learned about one, you have all the skills you’ll need to understand many more!
What is hypothesis testing?
Hypothesis Testing is a framework for asking questions about a dataset and answering them with probabilistics statements. There are many different kinds of hypothesis test that can be used to address defferent kinds of questions and data. In this article, we'll walk though a simulation of a one sample t-test in order to build an intuition about how many different kinds of hypothesis tests work.