Hypothesis Testing with Z-Test: Significance Level and Rejection Region

Join over 2 million students who advanced their careers with 365 Data Science. Learn from instructors who have worked at Meta, Spotify, Google, IKEA, Netflix, and Coca-Cola and master Python, SQL, Excel, machine learning, data analysis, AI fundamentals, and more.

what is a hypothesis test rejection region

If you want to understand why hypothesis testing works, you should first have an idea about the significance level and the reject region . We assume you already know what a hypothesis is , so let’s jump right into the action.

What Is the Significance Level?

First, we must define the term significance level .

Normally, we aim to reject the null if it is false.

Significance level

However, as with any test, there is a small chance that we could get it wrong and reject a null hypothesis that is true.

Error, significance level

How Is the Significance Level Denoted?

The significance level is denoted by α and is the probability of rejecting the null hypothesis , if it is true.

α and is the probability of rejecting the null hypothesis, significance level

So, the probability of making this error.

Typical values for α are 0.01, 0.05 and 0.1. It is a value that we select based on the certainty we need. In most cases, the choice of α is determined by the context we are operating in, but 0.05 is the most commonly used value.

Most common, significance level

A Case in Point

Say, we need to test if a machine is working properly. We would expect the test to make little or no mistakes. As we want to be very precise, we should pick a low significance level such as 0.01.

The famous Coca Cola glass bottle is 12 ounces. If the machine pours 12.1 ounces, some of the liquid would be spilled, and the label would be damaged as well. So, in certain situations, we need to be as accurate as possible.

Significance level: Coca Cola example

Higher Degree of Error

However, if we are analyzing humans or companies, we would expect more random or at least uncertain behavior. Hence, a higher degree of error.

You expect more random behavior, significance level

For instance, if we want to predict how much Coca Cola its consumers drink on average, the difference between 12 ounces and 12.1 ounces will not be that crucial. So, we can choose a higher significance level like 0.05 or 0.1.

The difference between 12 and 12.1, significance level

Hypothesis Testing: Performing a Z-Test

Now that we have an idea about the significance level , let’s get to the mechanics of hypothesis testing.

Imagine you are consulting a university and want to carry out an analysis on how students are performing on average.

How students are performing on average, significance-level

The university dean believes that on average students have a GPA of 70%. Being the data-driven researcher that you are, you can’t simply agree with his opinion, so you start testing.

The null hypothesis is: The population mean grade is 70%.

This is a hypothesized value.

The alternative hypothesis is: The population mean grade is not 70%. You can see how both of them are denoted, below.

University Dean example: Null hypothesis equals the population mean

Visualizing the Grades

Assuming that the population of grades is normally distributed, all grades received by students should look in the following way.

Distribution of grades, significance level

That is the true population mean .

Performing a Z-test

Now, a test we would normally perform is the Z-test . The formula is:

Z equals the sample mean , minus the hypothesized mean , divided by the standard error .

Z equals the sample mean, minus the hypothesized mean, divided by the standard error, significance level

The idea is the following.

We are standardizing or scaling the sample mean we got. (You can quickly obtain it with our Mean, Median, Mode calculator .) If the sample mean is close enough to the hypothesized mean , then Z will be close to 0. Otherwise, it will be far away from it. Naturally, if the sample mean is exactly equal to the hypothesized mean , Z will be 0.

If the sample mean is exactly equal to the hypothesized mean, Z will be 0, significance level

In all these cases, we would accept the null hypothesis .

What Is the Rejection Region?

The question here is the following:

How big should Z be for us to reject the null hypothesis ?

Well, there is a cut-off line. Since we are conducting a two-sided or a two-tailed test, there are two cut-off lines, one on each side.

Distribution of Z (standard normal distribution), significance level

When we calculate Z , we will get a value. If this value falls into the middle part, then we cannot reject the null. If it falls outside, in the shaded region, then we reject the null hypothesis .

That is why the shaded part is called: rejection region , as you can see below.

Rejection region, significance level

What Does the Rejection Region Depend on?

The area that is cut-off actually depends on the significance level .

Say the level of significance , α , is 0.05. Then we have α divided by 2, or 0.025 on the left side and 0.025 on the right side.

The level of significance, α, is 0.05. Then we have α divided by 2, or 0.025 on the left side and 0.025 on the right side

Now these are values we can check from the z-table . When α is 0.025, Z is 1.96. So, 1.96 on the right side and minus 1.96 on the left side.

Therefore, if the value we get for Z from the test is lower than minus 1.96, or higher than 1.96, we will reject the null hypothesis . Otherwise, we will accept it.

One-sided test: Z score is 1.96

That’s more or less how hypothesis testing works.

We scale the sample mean with respect to the hypothesized value. If Z is close to 0, then we cannot reject the null. If it is far away from 0, then we reject the null hypothesis .

How does hypothesis testing work?

Example of One Tailed Test

What about one-sided tests? We have those too!

Let’s consider the following situation.

Paul says data scientists earn more than $125,000. So, H 0 is: μ 0 is bigger than $125,000.

The alternative is that μ 0 is lower or equal to 125,000.

Using the same significance level , this time, the whole rejection region is on the left. So, the rejection region has an area of α . Looking at the z-table, that corresponds to a Z -score of 1.645. Since it is on the left, it is with a minus sign.

One-sided test: Z score is 1.645

Accept or Reject

Now, when calculating our test statistic Z , if we get a value lower than -1.645, we would reject the null hypothesis . We do that because we have statistical evidence that the data scientist salary is less than $125,000. Otherwise, we would accept it.

One-sided test: Z score is - 1.645 - rejecting null hypothesis

Another One-Tailed Test

To exhaust all possibilities, let’s explore another one-tailed test.

Say the university dean told you that the average GPA students get is lower than 70%. In that case, the null hypothesis is:

μ 0 is lower than 70%.

While the alternative is:

μ 0` is bigger or equal to 70%.

University Dean example: Null hypothesis lower than the population mean

In this situation, the rejection region is on the right side. So, if the test statistic is bigger than the cut-off z-score, we would reject the null, otherwise, we wouldn’t.

One-sided test: test statistic is bigger than the cut-off z-score - reject the null hypothesis

Importance of the Significance Level and the Rejection Region

To sum up, the significance level and the reject region are quite crucial in the process of hypothesis testing. The level of significance conducts the accuracy of prediction. We (the researchers) choose it depending on how big of a difference a possible error could make. On the other hand, the reject region helps us decide whether or not to reject the null hypothesis . After reading this and putting both of them into use, you will realize how convenient they make your work.

Interested in taking your skills from good to great? Try statistics course for free !

Next Tutorial:  Providing a Few Linear Regression Examples

World-Class

Data Science

Learn with instructors from:

Iliya Valchanov

Co-founder of 365 Data Science

Iliya is a finance graduate with a strong quantitative background who chose the exciting path of a startup entrepreneur. He demonstrated a formidable affinity for numbers during his childhood, winning more than 90 national and international awards and competitions through the years. Iliya started teaching at university, helping other students learn statistics and econometrics. Inspired by his first happy students, he co-founded 365 Data Science to continue spreading knowledge. He authored several of the program’s online courses in mathematics, statistics, machine learning, and deep learning.

We Think you'll also like

Hypothesis Testing: Null Hypothesis and Alternative Hypothesis

Statistics Tutorials

Hypothesis Testing: Null Hypothesis and Alternative Hypothesis

Article by Iliya Valchanov

False Positive vs. False Negative: Type I and Type II Errors in Statistical Hypothesis Testing

Calculating and Using Covariance and Linear Correlation Coefficient

Calculating and Using Covariance and Linear Correlation Coefficient

Examples of Numerical and Categorical Variables

Examples of Numerical and Categorical Variables

LEARN STATISTICS EASILY

LEARN STATISTICS EASILY

Learn Data Analysis Now!

LEARN STATISTICS EASILY LOGO 2

What is: Rejection Region

What is: rejection region.

The term “Rejection Region” refers to a specific area in statistical hypothesis testing that determines whether to reject the null hypothesis. In the context of a statistical test, the rejection region is defined based on the significance level, often denoted by alpha (α), which represents the probability of making a Type I error. This area is critical in making informed decisions based on sample data, as it delineates the values of the test statistic that would lead to the rejection of the null hypothesis in favor of the alternative hypothesis.

 width=

Ad description. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

Understanding Hypothesis Testing

Hypothesis testing is a fundamental aspect of inferential statistics, allowing researchers to make conclusions about a population based on sample data. In this framework, two competing hypotheses are formulated: the null hypothesis (H0) and the alternative hypothesis (H1). The null hypothesis typically posits that there is no effect or no difference, while the alternative hypothesis suggests that there is an effect or a difference. The rejection region plays a pivotal role in this process, as it is the threshold that dictates the outcome of the hypothesis test.

Defining the Rejection Region

The rejection region is established by the chosen significance level (α), which is commonly set at 0.05, 0.01, or 0.10. This level indicates the probability of rejecting the null hypothesis when it is actually true. For a one-tailed test, the rejection region is located in one tail of the distribution, while for a two-tailed test, it is split between both tails. The exact boundaries of the rejection region are determined by the critical values associated with the test statistic, which are derived from the sampling distribution under the null hypothesis.

Types of Tests and Their Rejection Regions

Different statistical tests have distinct rejection regions based on their underlying distributions. For instance, in a z-test, the rejection region is defined using the standard normal distribution, while in a t-test, it is based on the t-distribution. The choice of test affects the shape and size of the rejection region, which in turn influences the likelihood of rejecting the null hypothesis. Understanding these differences is crucial for selecting the appropriate test and accurately interpreting the results.

Visualizing the Rejection Region

Visual representations of the rejection region can greatly enhance comprehension. Typically, a graph will depict the probability distribution of the test statistic, with the rejection region shaded to indicate the area where the null hypothesis would be rejected. This visualization helps to clarify the relationship between the significance level, the critical values, and the corresponding probabilities. By observing the graph, researchers can better grasp the implications of their findings and the likelihood of making errors in hypothesis testing.

Impact of Sample Size on the Rejection Region

The size of the sample used in hypothesis testing can significantly affect the rejection region. Larger sample sizes tend to provide more accurate estimates of population parameters, resulting in narrower confidence intervals and more precise critical values. Consequently, with larger samples, the rejection region may become more sensitive, increasing the likelihood of detecting true effects. Conversely, smaller sample sizes may lead to wider confidence intervals and a broader rejection region, potentially obscuring significant findings.

Type I and Type II Errors

Understanding the rejection region is essential for grasping the concepts of Type I and Type II errors. A Type I error occurs when the null hypothesis is incorrectly rejected, while a Type II error happens when the null hypothesis is not rejected when it should be. The rejection region is directly linked to Type I errors, as it defines the threshold for making such an error. Conversely, the probability of a Type II error (denoted as β) is influenced by the power of the test, which is the probability of correctly rejecting the null hypothesis when it is false.

Applications of Rejection Regions in Data Science

In data science, the concept of the rejection region is applied across various domains, including A/B testing, quality control, and predictive modeling. For instance, in A/B testing, businesses may use rejection regions to determine whether a new product feature significantly improves user engagement compared to the existing version. By establishing a clear rejection region, data scientists can make data-driven decisions that enhance product development and marketing strategies.

The rejection region is a fundamental concept in statistical hypothesis testing, providing a framework for decision-making based on sample data. By understanding its definition, implications, and applications, researchers and data scientists can effectively navigate the complexities of statistical analysis and draw meaningful conclusions from their findings.

what is a hypothesis test rejection region

what is a hypothesis test rejection region

  • The Open University
  • Accessibility hub
  • Guest user / Sign out
  • Study with The Open University

My OpenLearn Profile

Personalise your OpenLearn profile, save your favourite content and get recognition for your learning

About this free course

Become an ou student, download this course, share this free course.

Data analysis: hypothesis testing

Start this free course now. Just create an account and sign in. Enrol and complete the course for a free statement of participation or digital badge if available.

5.1 Acceptance and rejection regions

The z-score table you created in Activity 5 represents the area under the normal distribution bell curve left of z (as shown in Figure 17).

A symmetrical graph resembling a bell. Areas left of z are coloured orange in the graph.

A symmetrical graph resembling a bell. Areas left of z are coloured orange in the graph.

The entries in this table can be used to determine whether to accept or reject the null hypothesis.

Suppose a marketing team at a company wishes to test the hypothesis that a new ad campaign will lead to a significant increase in sales. The team could use a one-tailed test with the reject region in the upper (right) tail and an alpha level of 1%.

Using the table created in Activity 5, the team can identify the range of z-scores that correspond to this test. They can then calculate the test statistic based on the data collected from the sales during and after the ad campaign. If the calculated test statistic falls within the rejection region identified by the table, the team can reject the null hypothesis and conclude that the ad campaign has had a significant impact on sales. This information can be used by the marketing team to justify the investment in the ad campaign and to make informed decisions about future marketing strategies.

In the context of the marketing team's hypothesis testing, the reject region for the one-tailed test with an alpha level of 1% corresponds to the range of z-scores that fall within the top 1% of the normal distribution. Conversely, the acceptable range refers to the range of z-scores that corresponds to the remaining 99% of the distribution to the left of z . Using the table created in Activity 5, the marketing team can identify the specific range of z-scores that correspond to the acceptable range and the reject region. Based on this table, the z-score of 2.33 corresponds to the upper limit of the acceptable range, as the area to the left of z = 2.33 represents approximately 99% of the area under the curve.

Therefore, if the team obtains a calculated z-score that is greater than 2.33, they can reject the null hypothesis and conclude that the new ad campaign has had a significant impact on sales. This information can help the marketing team make data-driven decisions about future campaigns and allocate resources effectively to maximise sales and profits. Figure 18 below illustrates this.

A symmetrical graph reminiscent of a bell showing the z-score azis and the rejection regions of null hypothesis

A symmetrical graph reminiscent of a bell. The graph points out z-score axis. Areas left of z are coloured orange in the graph. It also circles the rejection regions of null hypothesis when z  = 2.33 and alpha = 0.01.

Other than creating a z-score table, you calculate the region to the left of z by using the Excel formula NORM.S.DIST(z, cumulative). For example, you can calculate the region left of z when z = 2.33 by simply entering 2.33 as a z-score and setting the cumulative to be ‘TRUE’ in this Excel formula.

A table showing the entry of Excel formula and value ‘NORM.S.DIST(2.33, TRUE)’

A picture of a table made in Excel. It shows the entry of Excel formula and value ‘NORM.S.DIST(2.33, TRUE)’.

A table displaying the result 0.9901

A picture of a table made in Excel. After the calculation, it displays the result (0.9901).

You should get a value reading of 0.9901, which is exactly what you found in the z-score table in row 2.3 and column 0.03.

Here is another question. If you want to test hypotheses using the two-tailed test with the alpha level equal to 0.5%, how can you determine the z-scores region to reject the null hypothesis?

The two-tailed test requires you to divide the levels of alpha by 2.

Therefore, α for the two-tailed test = 0.05/2 = 0.0250

As the z-score table shows the area to the left of the value of z, a two-tailed test requires you to identify two entries. The area of one entry covers 0.975 (97.5%) of the area (where 0.025 of the area is outside the value of z on the right tail), and the area of another entry covers 0.025 of the area on the left tail.

Using the z-score table, you can determine the z-score = 1.96 or -1.96. Therefore, you will reject the null hypothesis for obtained z-score > 1.96 or z-score < 1.96.

Previous

what is a hypothesis test rejection region

Hypothesis Testing for Means & Proportions

  •   1  
  • |   2  
  • |   3  
  • |   4  
  • |   5  
  • |   6  
  • |   7  
  • |   8  
  • |   9  
  • |   10  

On This Page sidebar

Hypothesis Testing: Upper-, Lower, and Two Tailed Tests

Type i and type ii errors.

Learn More sidebar

All Modules

More Resources sidebar

Z score Table

t score Table

The procedure for hypothesis testing is based on the ideas described above. Specifically, we set up competing hypotheses, select a random sample from the population of interest and compute summary statistics. We then determine whether the sample data supports the null or alternative hypotheses. The procedure can be broken down into the following five steps.  

  • Step 1. Set up hypotheses and select the level of significance α.

H 0 : Null hypothesis (no change, no difference);  

H 1 : Research hypothesis (investigator's belief); α =0.05

 

Upper-tailed, Lower-tailed, Two-tailed Tests

The research or alternative hypothesis can take one of three forms. An investigator might believe that the parameter has increased, decreased or changed. For example, an investigator might hypothesize:  

: μ > μ , where μ is the comparator or null value (e.g., μ =191 in our example about weight in men in 2006) and an increase is hypothesized - this type of test is called an ; : μ < μ , where a decrease is hypothesized and this is called a ; or : μ ≠ μ where a difference is hypothesized and this is called a .  

The exact form of the research hypothesis depends on the investigator's belief about the parameter of interest and whether it has possibly increased, decreased or is different from the null value. The research hypothesis is set up by the investigator before any data are collected.

 

  • Step 2. Select the appropriate test statistic.  

The test statistic is a single number that summarizes the sample information.   An example of a test statistic is the Z statistic computed as follows:

When the sample size is small, we will use t statistics (just as we did when constructing confidence intervals for small samples). As we present each scenario, alternative test statistics are provided along with conditions for their appropriate use.

  • Step 3.  Set up decision rule.  

The decision rule is a statement that tells under what circumstances to reject the null hypothesis. The decision rule is based on specific values of the test statistic (e.g., reject H 0 if Z > 1.645). The decision rule for a specific test depends on 3 factors: the research or alternative hypothesis, the test statistic and the level of significance. Each is discussed below.

  • The decision rule depends on whether an upper-tailed, lower-tailed, or two-tailed test is proposed. In an upper-tailed test the decision rule has investigators reject H 0 if the test statistic is larger than the critical value. In a lower-tailed test the decision rule has investigators reject H 0 if the test statistic is smaller than the critical value.  In a two-tailed test the decision rule has investigators reject H 0 if the test statistic is extreme, either larger than an upper critical value or smaller than a lower critical value.
  • The exact form of the test statistic is also important in determining the decision rule. If the test statistic follows the standard normal distribution (Z), then the decision rule will be based on the standard normal distribution. If the test statistic follows the t distribution, then the decision rule will be based on the t distribution. The appropriate critical value will be selected from the t distribution again depending on the specific alternative hypothesis and the level of significance.  
  • The third factor is the level of significance. The level of significance which is selected in Step 1 (e.g., α =0.05) dictates the critical value.   For example, in an upper tailed Z test, if α =0.05 then the critical value is Z=1.645.  

The following figures illustrate the rejection regions defined by the decision rule for upper-, lower- and two-tailed Z tests with α=0.05. Notice that the rejection regions are in the upper, lower and both tails of the curves, respectively. The decision rules are written below each figure.

Rejection Region for Upper-Tailed Z Test (H : μ > μ ) with α=0.05

The decision rule is: Reject H if Z 1.645.

 

 

α

Z

0.10

1.282

0.05

1.645

0.025

1.960

0.010

2.326

0.005

2.576

0.001

3.090

0.0001

3.719

Standard normal distribution with lower tail at -1.645 and alpha=0.05

Rejection Region for Lower-Tailed Z Test (H 1 : μ < μ 0 ) with α =0.05

The decision rule is: Reject H 0 if Z < 1.645.

a

Z

0.10

-1.282

0.05

-1.645

0.025

-1.960

0.010

-2.326

0.005

-2.576

0.001

-3.090

0.0001

-3.719

Standard normal distribution with two tails

Rejection Region for Two-Tailed Z Test (H 1 : μ ≠ μ 0 ) with α =0.05

The decision rule is: Reject H 0 if Z < -1.960 or if Z > 1.960.

0.20

1.282

0.10

1.645

0.05

1.960

0.010

2.576

0.001

3.291

0.0001

3.819

The complete table of critical values of Z for upper, lower and two-tailed tests can be found in the table of Z values to the right in "Other Resources."

Critical values of t for upper, lower and two-tailed tests can be found in the table of t values in "Other Resources."

  • Step 4. Compute the test statistic.  

Here we compute the test statistic by substituting the observed sample data into the test statistic identified in Step 2.

  • Step 5. Conclusion.  

The final conclusion is made by comparing the test statistic (which is a summary of the information observed in the sample) to the decision rule. The final conclusion will be either to reject the null hypothesis (because the sample data are very unlikely if the null hypothesis is true) or not to reject the null hypothesis (because the sample data are not very unlikely).  

If the null hypothesis is rejected, then an exact significance level is computed to describe the likelihood of observing the sample data assuming that the null hypothesis is true. The exact level of significance is called the p-value and it will be less than the chosen level of significance if we reject H 0 .

Statistical computing packages provide exact p-values as part of their standard output for hypothesis tests. In fact, when using a statistical computing package, the steps outlined about can be abbreviated. The hypotheses (step 1) should always be set up in advance of any analysis and the significance criterion should also be determined (e.g., α =0.05). Statistical computing packages will produce the test statistic (usually reporting the test statistic as t) and a p-value. The investigator can then determine statistical significance using the following: If p < α then reject H 0 .  

 

 

  • Step 1. Set up hypotheses and determine level of significance

H 0 : μ = 191 H 1 : μ > 191                 α =0.05

The research hypothesis is that weights have increased, and therefore an upper tailed test is used.

  • Step 2. Select the appropriate test statistic.

Because the sample size is large (n > 30) the appropriate test statistic is

  • Step 3. Set up decision rule.  

In this example, we are performing an upper tailed test (H 1 : μ> 191), with a Z test statistic and selected α =0.05.   Reject H 0 if Z > 1.645.

We now substitute the sample data into the formula for the test statistic identified in Step 2.  

We reject H 0 because 2.38 > 1.645. We have statistically significant evidence at a =0.05, to show that the mean weight in men in 2006 is more than 191 pounds. Because we rejected the null hypothesis, we now approximate the p-value which is the likelihood of observing the sample data if the null hypothesis is true. An alternative definition of the p-value is the smallest level of significance where we can still reject H 0 . In this example, we observed Z=2.38 and for α=0.05, the critical value was 1.645. Because 2.38 exceeded 1.645 we rejected H 0 . In our conclusion we reported a statistically significant increase in mean weight at a 5% level of significance. Using the table of critical values for upper tailed tests, we can approximate the p-value. If we select α=0.025, the critical value is 1.96, and we still reject H 0 because 2.38 > 1.960. If we select α=0.010 the critical value is 2.326, and we still reject H 0 because 2.38 > 2.326. However, if we select α=0.005, the critical value is 2.576, and we cannot reject H 0 because 2.38 < 2.576. Therefore, the smallest α where we still reject H 0 is 0.010. This is the p-value. A statistical computing package would produce a more precise p-value which would be in between 0.005 and 0.010. Here we are approximating the p-value and would report p < 0.010.                  

In all tests of hypothesis, there are two types of errors that can be committed. The first is called a Type I error and refers to the situation where we incorrectly reject H 0 when in fact it is true. This is also called a false positive result (as we incorrectly conclude that the research hypothesis is true when in fact it is not). When we run a test of hypothesis and decide to reject H 0 (e.g., because the test statistic exceeds the critical value in an upper tailed test) then either we make a correct decision because the research hypothesis is true or we commit a Type I error. The different conclusions are summarized in the table below. Note that we will never know whether the null hypothesis is really true or false (i.e., we will never know which row of the following table reflects reality).

Table - Conclusions in Test of Hypothesis

 

is True

Correct Decision

Type I Error

is False

Type II Error

Correct Decision

In the first step of the hypothesis test, we select a level of significance, α, and α= P(Type I error). Because we purposely select a small value for α, we control the probability of committing a Type I error. For example, if we select α=0.05, and our test tells us to reject H 0 , then there is a 5% probability that we commit a Type I error. Most investigators are very comfortable with this and are confident when rejecting H 0 that the research hypothesis is true (as it is the more likely scenario when we reject H 0 ).

When we run a test of hypothesis and decide not to reject H 0 (e.g., because the test statistic is below the critical value in an upper tailed test) then either we make a correct decision because the null hypothesis is true or we commit a Type II error. Beta (β) represents the probability of a Type II error and is defined as follows: β=P(Type II error) = P(Do not Reject H 0 | H 0 is false). Unfortunately, we cannot choose β to be small (e.g., 0.05) to control the probability of committing a Type II error because β depends on several factors including the sample size, α, and the research hypothesis. When we do not reject H 0 , it may be very likely that we are committing a Type II error (i.e., failing to reject H 0 when in fact it is false). Therefore, when tests are run and the null hypothesis is not rejected we often make a weak concluding statement allowing for the possibility that we might be committing a Type II error. If we do not reject H 0 , we conclude that we do not have significant evidence to show that H 1 is true. We do not conclude that H 0 is true.

Lightbulb icon signifying an important idea

 The most common reason for a Type II error is a small sample size.

return to top | previous page | next page

Content ©2017. All Rights Reserved. Date last modified: November 6, 2017. Wayne W. LaMorte, MD, PhD, MPH

  • Skip to secondary menu
  • Skip to main content
  • Skip to primary sidebar

Statistics By Jim

Making statistics intuitive

Critical Value: Definition, Finding & Calculator

By Jim Frost 2 Comments

What is a Critical Value?

A critical value defines regions in the sampling distribution of a test statistic. These values play a role in both hypothesis tests and confidence intervals. In hypothesis tests, critical values determine whether the results are statistically significant. For confidence intervals, they help calculate the upper and lower limits.

In both cases, critical values account for uncertainty in sample data you’re using to make inferences about a population . They answer the following questions:

  • How different does the sample estimate need to be from the null hypothesis to be statistically significant?
  • What is the margin of error (confidence interval) around the sample estimate of the population parameter ?

In this post, I’ll show you how to find critical values, use them to determine statistical significance, and use them to construct confidence intervals. I also include a critical value calculator at the end of this article so you can apply what you learn.

Because most people start learning with the z-test and its test statistic, the z-score, I’ll use them for the examples throughout this post. However, I provide links with detailed information for other types of tests and sampling distributions.

Related posts : Sampling Distributions and Test Statistics

Using a Critical Value to Determine Statistical Significance

Diagram showing critical region in a distribution.

In this context, the sampling distribution of a test statistic defines the probability for ranges of values. The significance level (α) specifies the probability that corresponds with the critical value within the distribution. Let’s work through an example for a z-test.

The z-test uses the z test statistic. For this test, the z-distribution finds probabilities for ranges of z-scores under the assumption that the null hypothesis is true. For a z-test, the null z-score is zero, which is at the central peak of the sampling distribution. This sampling distribution centers on the null hypothesis value, and the critical values mark the minimum distance from the null hypothesis required for statistical significance.

Critical values depend on your significance level and whether you’re performing a one- or two-sided hypothesis. For these examples, I’ll use a significance level of 0.05. This value defines how improbable the test statistic must be to be significant.

Related posts : Significance Levels and P-values and Z-scores

Two-Sided Tests

Two-sided hypothesis tests have two rejection regions. Consequently, you’ll need two critical values that define them. Because there are two rejection regions, we must split our significance level in half. Each rejection region has a probability of α / 2, making the total likelihood for both areas equal the significance level.

The probability plot below displays the critical values and the rejection regions for a two-sided z-test with a significance level of 0.05. When the z-score is ≤ -1.96 or ≥ 1.96, it exceeds the cutoff, and your results are statistically significant.

Graph that displays critical values for a two-sided test.

One-Sided Tests

One-tailed tests have one rejection region and, hence, only one critical value. The total α probability goes into that one side. The probability plots below display these values for right- and left-sided z-tests. These tests can detect effects in only one direction.

Graph that displays a critical value for a right-sided test.

Related post : Understanding One-Tailed and Two-Tailed Hypothesis Tests and Effects in Statistics

Using a Critical Value to Construct Confidence Intervals

Confidence intervals use the same critical values (CVs) as the corresponding hypothesis test. The confidence level equals 1 – the significance level. Consequently, the CVs for a significance level of 0.05 produce a confidence level of 1 – 0.05 = 0.95 or 95%.

For example, to calculate the 95% confidence interval for our two-tailed z-test with a significance level of 0.05, use the CVs of -1.96 and 1.96 that we found above.

To calculate the upper and lower limits of the interval, take the positive critical value and multiply it by the standard error of the mean. Then take the sample mean and add and subtract that product from it.

  • Lower Limit = Sample Mean – (CV * Standard Error of the Mean)
  • Upper Limit = Sample Mean + (CV * Standard Error of the Mean)

To learn more about confidence intervals and how to construct them, read my posts about Confidence Intervals and How Confidence Intervals Work .

Related post : Standard Error of the Mean

How to Find a Critical Value

Unfortunately, the formulas for finding critical values are very complex. Typically, you don’t calculate them by hand. For the examples in this article, I’ve used statistical software to find them. However, you can also use statistical tables.

To learn how to use these critical value tables, read my articles that contain the tables and information about using them. The process for finding them is similar for the various tests. Using these tables requires knowing the correct test statistic, the significance level, the number of tails, and, in most cases, the degrees of freedom.

The following articles provide the statistical tables, explain how to use them, and visually illustrate the results.

  • T distribution table
  • Chi-square table

Related post : Degrees of Freedom

Critical Value Calculator

Another method for finding CVs is to use a critical value calculator, such as the one below. These calculators are handy for finding the answer, but they don’t provide the context for the results.

This calculator finds critical values for the sampling distributions of common test statistics.

For example, choose the following in the calculator:

  • Z (standard normal)
  • Significance level = 0.05

The calculator will display the same ±1.96 values we found earlier in this article.

Share this:

what is a hypothesis test rejection region

Reader Interactions

' src=

January 16, 2024 at 5:26 pm

Hello, I am currently taking statistics and am reviewing confidence intervals. I would like to know what is the equation for calculating a two-tailed test for upper and lower limits? I would like to know is there a way to calculate one and two-tailed tests without using a confidence interval calculator and can you explain further?

' src=

January 16, 2024 at 6:43 pm

If you’re talking about calculating the critical values values for a test statistic for two-tailed test, the calculations are fairly complex. Consequently, you’ll either use statistical software, an online calculator, or a statistical table to find those limits.

Comments and Questions Cancel reply

Acceptance Region: Simple Definition & Example

Hypothesis Testing >

acceptance region

More Formal Definition of Acceptance Region

According to the The Concise Encyclopedia of Statistics , the acceptance region is “…the interval within the sampling distribution of the test statistic that is consistent with the null hypothesis H 0 from hypothesis testing .” In more simple terms, let’s say you run a hypothesis test like a z-test . The results of the test come in the form of a z-value , which has a large range of possible values. Within that range of values, some will fall into an interval that suggests the null hypothesis is correct. That interval is the acceptance region.

Why is it “Provisionally” Acceptance?

You provisionally accept the null hypothesis because a hypothesis test doesn’t tell you which hypothesis is true (the null or alternate hypothesis ), or even which is probably true. The only thing is tests is whether there’s enough evidence in your data to reject the null hypothesis. Failure to accept the alternate hypothesis doesn’t make the null hypothesis true.

Let’s say our “experiment” is where you caught a child red-handed with a stolen cookie:

  • Null hypothesis (H 0 ): The child didn’t steal the cookie (innocent until proven guilty!).
  • Alternate hypothesis (H 1 ): The child did steal the cookie.

You’re pretty certain the child stole the cookie. But after gathering all evidence, you don’t find enough evidence to say for sure that the child is guilty. Therefore, there isn’t enough evidence in support of the alternate hypothesis that the child is guilty . In other words, you can’t reject the null hypothesis that the child is innocent in favor of the hypothesis that the child is guilty. That doesn’t mean the child is innocent. You just didn’t have enough evidence to prove them guilty. Although your result fell into the acceptance region, you don’t actually “accept” the null hypothesis of innocence. You just provisionally (perhaps begrudgingly) accept it and let the child off without punishment. Later on you might find crumbs in their bed, leading you to revisit your findings.

This subtle difference may seem pedantic, and in elementary statistics it’s usually not an important matter to stress. However, if you plan to publish your results you should never say you “accept the null hypothesis”. You can say that you provisionally accept it, or that you failed to reject it . Check with your intended publication (or with your professor) to see what wording they prefer.

Dodge, Y. (2008). The Concise Encyclopedia of Statistics . Springer. Gonick, L. (1993). The Cartoon Guide to Statistics . HarperPerennial.

Leave a Comment

You must be logged in to post a comment.

Pardon Our Interruption

As you were browsing something about your browser made us think you were a bot. There are a few reasons this might happen:

  • You've disabled JavaScript in your web browser.
  • You're a power user moving through this website with super-human speed.
  • You've disabled cookies in your web browser.
  • A third-party browser plugin, such as Ghostery or NoScript, is preventing JavaScript from running. Additional information is available in this support article .

To regain access, please make sure that cookies and JavaScript are enabled before reloading the page.

what is a hypothesis test rejection region

User Preferences

Content preview.

Arcu felis bibendum ut tristique et egestas quis:

  • Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris
  • Duis aute irure dolor in reprehenderit in voluptate
  • Excepteur sint occaecat cupidatat non proident

Keyboard Shortcuts

6a.4.2 - more on the p-value and rejection region approach, two methods for making a statistical decision section  .

Of the two methods for making a statistical decision, the p-value approach is more commonly used and provided in published literature. However, understanding the rejection region approach can go a long way in one's understanding of the p-value method. In the video, we show how the two methods are related. Regardless of the method applied, the conclusions from the two approaches are exactly the same.

Video: The Rejection Region vs the P-Value Approach

Comparing the Two Approaches Section  

Both approaches will ensure the same conclusion and either one will work. However, using the p-value approach has the following advantages:

  • Using the rejection region approach, you need to check the table or software for the critical value every time you use a different \(\alpha \) value.
  • In addition to just using it to reject or not reject \(H_0 \) by comparing p-value to \(\alpha \) value, the p-value also gives us some idea of the strength of the evidence against \(H_0 \).

Stack Exchange Network

Stack Exchange network consists of 183 Q&A communities including Stack Overflow , the largest, most trusted online community for developers to learn, share their knowledge, and build their careers.

Q&A for work

Connect and share knowledge within a single location that is structured and easy to search.

Different ways of determining rejection region of a two sided test

In a two sided test, assume the test statistic has been chosen to be $T(X)$ and the distribution of $T(X)$ under null hypothesis is also known to be $F$. Let the significance level be $\alpha$.

I can come up with two different ways to determine the rejection region:

$\{|T(x) - \mu| > c\}$. $\mu$ is the mean of the distribution $F$ of $T(X)$under null, and $c$ is determined by solving $$\inf_{c \geq 0} c$$ subject to $$P_{T(X) \sim F} (|T(X) - \mu| > c) \leq \alpha.$$ So the rejection region is symmetric around $\mu$.

$\{T(x) > c_1\} \cup \{T(x) < c_2\}$. $c_1$ and $c_2$ are determined by solving $$\inf_{c_1 \in \mathbb R} c_1$$ subject to $$P_{T(X) \sim F} (T(X) > c_1) \leq \alpha/2$$ and $$\sup_{c_2 \in \mathbb R} c_2$$ subject to $$P_{T(X) \sim F} (T(X) < c_2) \leq \alpha/2.$$ So the rejection region evenly split $\alpha$ to both sides.

Am I correct that those two methods will agree when the null distribution $F$of $T(X)$ is symmetric around its mean $\mu$, and may not agree when $F$ isn't symmetric around $\mu$?

I was wondering what advantage and disadvantages these two methods have? Which one is recommended and when?

Are both methods used in some textbooks? If yes, references?

What are some other methods for two-sided tests? For example, can we generalized the second method by splitting $\alpha$ arbitrarily unevenly to the two sides?

Consider the relation between rejection region in testing and confidence interval. Are the above discussions also apply to confidence intervals?

Thanks and regards!

  • hypothesis-testing

Tim's user avatar

Question 1. Right, if $F$ is symmetric (and continuous: weird things may happen if some points have non-zero weight and you cut your interval right on them) you have $P(T(X)-\mu) > c = P(T(X)-\mu < -c)$, so you can split the rejection region into two unbounded intervals, each of probability $\alpha / 2$.

Questions 2,4. For any set $C$, as long as $P(T(X) \in C) = \alpha$, you're fine. You could even choose a region such as $[\mu-d,\mu + d]$ ! It is however not natural, because one would like your rejection set to include weird values of the statistic (values that are far from $\mu$) rather than normal values. Why? Because the test statistic is expected to be a measure of the distance between the data and the null hypothesis. For instance, in the lady tasting tea, the test statistic is the number of correctly classified cups. One would expect that the higher it is, the more the ability of the lady is proven. A weird rejection set could be designed, that could include the lady correctly classifying all cups, but not the lady correctly classifying all cups but one. It would be perfectly acceptable from a formal point of vue.

That insisting on having centered confidence intervals is not based on mathematics strictly can be seen easily: if $[a,b]$ is a centered confidence interval for parameter $\theta$, $[f(a),f(b)]$ has no reason to be a centered confidence interval for transformed parameter $f(\theta)$. For instance, say your data is sound power, and that you want to relate that to medical damages or sound perception; do you want intervals centered on the original scale ($W.m^{-2}$ for instance), or on the transformed log scale (dB) ?

pkr's user avatar

Your Answer

Sign up or log in, post as a guest.

Required, but never shown

By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy .

Not the answer you're looking for? Browse other questions tagged hypothesis-testing or ask your own question .

  • Featured on Meta
  • We've made changes to our Terms of Service & Privacy Policy - July 2024
  • Bringing clarity to status tag usage on meta sites

Hot Network Questions

  • Can light become a satellite of a black hole?
  • Trying to find an old book (fantasy or scifi?) in which the protagonist and their romantic partner live in opposite directions in time
  • How to make a ParametricPlot3D into solid shape
  • "can-do" vs. "can-explain" fallacy
  • How is the grammar of this sentence explained?
  • Soft links on Python
  • Do mini-humans need a "real" Saturn V to reach the moon?
  • Clarification about a notation (or typo?) in ComplexityZoo for QAM class
  • Is this an accurate interpretation of the elasticity magus arcana?
  • What are the limits of Terms of Service as a legal shield for a company?
  • How to prevent nicematrix from intersecting brackets of submatrices?
  • Form Builder: Create Contact and add to (static) group
  • What's the origin of the colloquial "peachy", "simply peachy", and "just peachy"?
  • Canceling factors in a ratio of factorials
  • Is it possible to create a board position where White must make the move that leads to stalemating Black to avoid Black stalemating White?
  • Submitting a paper as a nonacademic practitioner in a field
  • using a tikz foreach loop inside a newcommand
  • Reduce String Length With Thread Safety & Concurrency
  • Electric skateboard: helmet replacement
  • Solve an equation perturbatively
  • Seth and Cain take turns picking numbers from 1 to 50. Who wins?
  • Why do we reduce a body to its center of mass when calculating gain/loss of gravitational potential energy?
  • When a submarine blows its ballast and rises, where did the energy for the ascent come from?
  • How many kms do aluminum rims last when ridden in city and/or road conditions?

what is a hypothesis test rejection region

IMAGES

  1. PPT

    what is a hypothesis test rejection region

  2. Hypothesis Testing and Confidence Intervals

    what is a hypothesis test rejection region

  3. PPT

    what is a hypothesis test rejection region

  4. Rejection Region in Hypothesis Testing » Data Science Tutorials

    what is a hypothesis test rejection region

  5. T Test (Students T Test)

    what is a hypothesis test rejection region

  6. PPT

    what is a hypothesis test rejection region

COMMENTS

  1. Rejection Region Definition Statistics How To

    A one tailed test with the rejection region in one tail. Rejection Regions and P-Values. There are two ways you can test a hypothesis: with a p-value and with a critical value. P-value method: When you run a hypothesis test (for example, a z test), the result of that test will be a p value. The p value is a "probability value."

  2. Hypothesis Testing: Significance Level & Rejection Region

    In this situation, the rejection region is on the right side. So, if the test statistic is bigger than the cut-off z-score, we would reject the null, otherwise, we wouldn't. Importance of the Significance Level and the Rejection Region. To sum up, the significance level and the reject region are quite crucial in the process of hypothesis ...

  3. 7.5: Critical values, p-values, and significance level

    Figure \(\PageIndex{1}\): The rejection region for a one-tailed test. The shaded rejection region takes us 5% of the area under the curve. Any result which falls in that region is sufficient evidence to reject the null hypothesis. The rejection region is bounded by a specific \(z\)-value, as is any area under the curve.

  4. 6a.2

    In testing a hypothesis, we use a method where we gather data in an effort to gather evidence about the hypothesis. ... The rejection region is found by using alpha to find a critical value; the rejection region is the area that is more extreme than the critical value. We discuss the p-value and rejection region in more detail in the next section.

  5. 6a.4.1

    The rejection region is the region where, if our test statistic falls, then we have enough evidence to reject the null hypothesis. If we consider the right-tailed test, for example, the rejection region is any value greater than \(c_{1-\alpha} \), where \(c_{1-\alpha}\) is the critical value.

  6. S.3.1 Hypothesis Testing (Critical Value Approach)

    The critical value for conducting the right-tailed test H0 : μ = 3 versus HA : μ > 3 is the t -value, denoted t\ (\alpha\), n - 1, such that the probability to the right of it is \ (\alpha\). It can be shown using either statistical software or a t -table that the critical value t 0.05,14 is 1.7613. That is, we would reject the null ...

  7. What is: Rejection Region

    The rejection region is a fundamental concept in statistical hypothesis testing, providing a framework for decision-making based on sample data. By understanding its definition, implications, and applications, researchers and data scientists can effectively navigate the complexities of statistical analysis and draw meaningful conclusions from ...

  8. 9.1: Introduction to Hypothesis Testing

    In hypothesis testing, the goal is to see if there is sufficient statistical evidence to reject a presumed null hypothesis in favor of a conjectured alternative hypothesis.The null hypothesis is usually denoted \(H_0\) while the alternative hypothesis is usually denoted \(H_1\). An hypothesis test is a statistical decision; the conclusion will either be to reject the null hypothesis in favor ...

  9. Data analysis: hypothesis testing: 5.1 Acceptance and rejection regions

    The team could use a one-tailed test with the reject region in the upper (right) tail and an alpha level of 1%. ... In the context of the marketing team's hypothesis testing, the reject region for the one-tailed test with an alpha level of 1% corresponds to the range of z-scores that fall within the top 1% of the normal distribution.

  10. Hypothesis Testing: Upper-, Lower, and Two Tailed Tests

    The decision rule is a statement that tells under what circumstances to reject the null hypothesis. The decision rule is based on specific values of the test statistic (e.g., reject H 0 if Z > 1.645). The decision rule for a specific test depends on 3 factors: the research or alternative hypothesis, the test statistic and the level of significance.

  11. Statistical hypothesis test

    Region of rejection / Critical region: The set of values of the test statistic for which the null hypothesis is rejected. Power of a test (1 − β ) Size : For simple hypotheses, this is the test's probability of incorrectly rejecting the null hypothesis.

  12. 8.1: The Elements of Hypothesis Testing

    Hypothesis testing is a statistical procedure in which a choice is made between a null hypothesis and an alternative hypothesis based on information in a sample. The end result of a hypotheses testing procedure is a choice of one of the following two possible conclusions: Reject H0. H 0. (and therefore accept Ha.

  13. Hypothesis Testing

    In this hypothesis testing video we discuss how to find rejection regions and critical values using a z test, when the standard deviation is known. We cover...

  14. 6a.4.3

    Step 1: Set up the hypotheses and check conditions. Step 2: Decide on the level of significance \ (\boldsymbol { (\alpha)}\). Step 3: Calculate the test statistic. The first few steps (Step 1 - Step 3) are exactly the same as the rejection region or p-value approach.

  15. Critical Value: Definition, Finding & Calculator

    Test statistics that exceed a critical value have a low probability of occurring if the null hypothesis is true. Therefore, when test statistics exceed these cutoffs, you can reject the null and conclude that the effect exists in the population. In other words, they define the rejection regions for the null hypothesis.

  16. Lesson 10

    This is just a few minutes of a complete course. Get full lessons & more subjects at: http://www.MathTutorDVD.com.

  17. PDF Hypothesis testing

    Hypothesis Testing Using Rejection Regions. Tip: The only differences in procedure between using Rejection Regions and P-values are the steps in the two boxes that have a doubled border. The rest of the procedure is identical. Compare the standardized test statistic with the critical value(s).

  18. hypothesis testing

    The test statistic and the p-value are calculated after collecting the data under the assumption the null hypothesis is true. Different samples will result in different test statistics and p-values, but the rejection region and significance level will not change. You may formulate your decision rule in one of two equivalent ways.

  19. Acceptance Region: Simple Definition & Example

    Hypothesis Testing >. Results from a statistical tests will fall into one of two regions: the rejection region— which will lead you to reject the null hypothesis, or the acceptance region, where you provisionally accept the null hypothesis. The acceptance region is basically the complement of the rejection region; If your result does not fall into the rejection region, it must fall into the ...

  20. 1.6

    This region, which leads to rejection of the null hypothesis, is called the rejection region. For example, for a significance level of 5%: For an upper-tail test, the critical value is the 95th percentile of the t-distribution with n−1 degrees of freedom; reject the null in favor of the alternative if the t-statistic is greater than this.

  21. Mastering Hypothesis Testing: 8 Steps Decoded

    My test statistic is t = -2.548 and my CV = -1.649. .This would be: -2.548 < -1.649 This means that the test statistic falls in the critical region also known as the rejection region and the null hypothesis should be rejected. Since I reject the null hypothesis, it means that the average salary for all jobs in Minnesota does not equal

  22. hypothesis testing

    6. The rejection region is fixed beforehand. If the null hypothesis is true then some α% α % of the observations will be in the region. The p-value is not the same as this α% α %. The p-value is computed for each separate observation, and can be different for two observations that both fall inside the rejection region.

  23. 6a.4.2

    Using the rejection region approach, you need to check the table or software for the critical value every time you use a different α value. In addition to just using it to reject or not reject H 0 by comparing p-value to α value, the p-value also gives us some idea of the strength of the evidence against H 0.

  24. hypothesis testing

    You could even choose a region such as $[\mu-d,\mu + d]$ ! It is however not natural, because one would like your rejection set to include weird values of the statistic (values that are far from $\mu$) rather than normal values. Why? Because the test statistic is expected to be a measure of the distance between the data and the null hypothesis ...