| α | Z | 0.10 | 1.282 | 0.05 | 1.645 | 0.025 | 1.960 | 0.010 | 2.326 | 0.005 | 2.576 | 0.001 | 3.090 | 0.0001 | 3.719 | Rejection Region for Lower-Tailed Z Test (H 1 : μ < μ 0 ) with α =0.05 The decision rule is: Reject H 0 if Z < 1.645. | a | Z | 0.10 | -1.282 | 0.05 | -1.645 | 0.025 | -1.960 | 0.010 | -2.326 | 0.005 | -2.576 | 0.001 | -3.090 | 0.0001 | -3.719 | Rejection Region for Two-Tailed Z Test (H 1 : μ ≠ μ 0 ) with α =0.05 The decision rule is: Reject H 0 if Z < -1.960 or if Z > 1.960. | | | 0.20 | 1.282 | 0.10 | 1.645 | 0.05 | 1.960 | 0.010 | 2.576 | 0.001 | 3.291 | 0.0001 | 3.819 | The complete table of critical values of Z for upper, lower and two-tailed tests can be found in the table of Z values to the right in "Other Resources." Critical values of t for upper, lower and two-tailed tests can be found in the table of t values in "Other Resources." - Step 4. Compute the test statistic.
Here we compute the test statistic by substituting the observed sample data into the test statistic identified in Step 2. The final conclusion is made by comparing the test statistic (which is a summary of the information observed in the sample) to the decision rule. The final conclusion will be either to reject the null hypothesis (because the sample data are very unlikely if the null hypothesis is true) or not to reject the null hypothesis (because the sample data are not very unlikely). If the null hypothesis is rejected, then an exact significance level is computed to describe the likelihood of observing the sample data assuming that the null hypothesis is true. The exact level of significance is called the p-value and it will be less than the chosen level of significance if we reject H 0 . Statistical computing packages provide exact p-values as part of their standard output for hypothesis tests. In fact, when using a statistical computing package, the steps outlined about can be abbreviated. The hypotheses (step 1) should always be set up in advance of any analysis and the significance criterion should also be determined (e.g., α =0.05). Statistical computing packages will produce the test statistic (usually reporting the test statistic as t) and a p-value. The investigator can then determine statistical significance using the following: If p < α then reject H 0 . - Step 1. Set up hypotheses and determine level of significance
H 0 : μ = 191 H 1 : μ > 191 α =0.05 The research hypothesis is that weights have increased, and therefore an upper tailed test is used. - Step 2. Select the appropriate test statistic.
Because the sample size is large (n > 30) the appropriate test statistic is - Step 3. Set up decision rule.
In this example, we are performing an upper tailed test (H 1 : μ> 191), with a Z test statistic and selected α =0.05. Reject H 0 if Z > 1.645. We now substitute the sample data into the formula for the test statistic identified in Step 2. We reject H 0 because 2.38 > 1.645. We have statistically significant evidence at a =0.05, to show that the mean weight in men in 2006 is more than 191 pounds. Because we rejected the null hypothesis, we now approximate the p-value which is the likelihood of observing the sample data if the null hypothesis is true. An alternative definition of the p-value is the smallest level of significance where we can still reject H 0 . In this example, we observed Z=2.38 and for α=0.05, the critical value was 1.645. Because 2.38 exceeded 1.645 we rejected H 0 . In our conclusion we reported a statistically significant increase in mean weight at a 5% level of significance. Using the table of critical values for upper tailed tests, we can approximate the p-value. If we select α=0.025, the critical value is 1.96, and we still reject H 0 because 2.38 > 1.960. If we select α=0.010 the critical value is 2.326, and we still reject H 0 because 2.38 > 2.326. However, if we select α=0.005, the critical value is 2.576, and we cannot reject H 0 because 2.38 < 2.576. Therefore, the smallest α where we still reject H 0 is 0.010. This is the p-value. A statistical computing package would produce a more precise p-value which would be in between 0.005 and 0.010. Here we are approximating the p-value and would report p < 0.010. In all tests of hypothesis, there are two types of errors that can be committed. The first is called a Type I error and refers to the situation where we incorrectly reject H 0 when in fact it is true. This is also called a false positive result (as we incorrectly conclude that the research hypothesis is true when in fact it is not). When we run a test of hypothesis and decide to reject H 0 (e.g., because the test statistic exceeds the critical value in an upper tailed test) then either we make a correct decision because the research hypothesis is true or we commit a Type I error. The different conclusions are summarized in the table below. Note that we will never know whether the null hypothesis is really true or false (i.e., we will never know which row of the following table reflects reality). Table - Conclusions in Test of Hypothesis | | | is True | Correct Decision | Type I Error | is False | Type II Error | Correct Decision | In the first step of the hypothesis test, we select a level of significance, α, and α= P(Type I error). Because we purposely select a small value for α, we control the probability of committing a Type I error. For example, if we select α=0.05, and our test tells us to reject H 0 , then there is a 5% probability that we commit a Type I error. Most investigators are very comfortable with this and are confident when rejecting H 0 that the research hypothesis is true (as it is the more likely scenario when we reject H 0 ). When we run a test of hypothesis and decide not to reject H 0 (e.g., because the test statistic is below the critical value in an upper tailed test) then either we make a correct decision because the null hypothesis is true or we commit a Type II error. Beta (β) represents the probability of a Type II error and is defined as follows: β=P(Type II error) = P(Do not Reject H 0 | H 0 is false). Unfortunately, we cannot choose β to be small (e.g., 0.05) to control the probability of committing a Type II error because β depends on several factors including the sample size, α, and the research hypothesis. When we do not reject H 0 , it may be very likely that we are committing a Type II error (i.e., failing to reject H 0 when in fact it is false). Therefore, when tests are run and the null hypothesis is not rejected we often make a weak concluding statement allowing for the possibility that we might be committing a Type II error. If we do not reject H 0 , we conclude that we do not have significant evidence to show that H 1 is true. We do not conclude that H 0 is true. The most common reason for a Type II error is a small sample size. return to top | previous page | next page Content ©2017. All Rights Reserved. Date last modified: November 6, 2017. Wayne W. LaMorte, MD, PhD, MPH Critical Region and Confidence IntervalContents Toggle Main Menu 1 Confidence Interval 2 Significance Levels 3 Critical Region 4 Critical Values 5 Constructing a Confidence Interval 5.1 Binomial Distribution 5.2 Normal Distribution 5.3 Student $t$-distribution 6 Video Examples Confidence IntervalA confidence interval , also known as the acceptance region, is a set of values for the test statistic for which the null hypothesis is accepted. i.e. if the observed test statistic is in the confidence interval then we accept the null hypothesis and reject the alternative hypothesis . Significance LevelsConfidence intervals can be calculated at different significance levels . We use $\alpha$ to denote the level of significance and perform a hypothesis test with a $100(1- \alpha)$% confidence interval. Confidence intervals are usually calculated at $5$% or $1$% significance levels, for which $\alpha = 0.05$ and $\alpha = 0.01$ respectively. Note that a $95$% confidence interval does not mean there is a $95$% chance that the true value being estimated is in the calculated interval. Rather, given a population, there is a $95$% chance that choosing a random sample from this population results in a confidence interval which contains the true value being estimated. Critical RegionA critical region , also known as the rejection region, is a set of values for the test statistic for which the null hypothesis is rejected. i.e. if the observed test statistic is in the critical region then we reject the null hypothesis and accept the alternative hypothesis. Critical ValuesThe critical value at a certain significance level can be thought of as a cut-off point. If a test statistic on one side of the critical value results in accepting the null hypothesis, a test statistic on the other side will result in rejecting the null hypothesis. Constructing a Confidence IntervalBinomial distribution. Usually, the easiest way to perform a hypothesis test with the binomial distribution is to use the $p$-value and see whether it is larger or smaller than $\alpha$, the significance level used. Sometimes, if we have observed a large number of Bernoulli Trials, we can use the observed probability of success $\hat{p}$, based entirely on the data obtained, to approximate the distribution of error using the normal distribution. We do this using the formula \[\hat{p} \pm z_{1-\frac{\alpha}{2} } \sqrt{ \frac{1}{n} \hat{p} (1-\hat {p})}\] where $\hat{p}$ is the estimated probability of success, $z_{1- \frac{\alpha}{2} }$ is obtained from the normal distribution tables , $\alpha$ is the significance level and $n$ is the sample size. Worked ExampleA coin is tossed $1050$ times and lands on heads $500$ times. Construct a $90$% confidence interval for the probability $p$ of getting a head. Here the observed probability of success $\hat{p} = \dfrac{500}{1050}$, $n=1050$ and $\alpha = 0.1$ so $z_{1-\frac{\alpha}{2} } = z_{0.95} = 1.645$. This is because $\Phi^{-1} (0.95) = 1.645$ . So the confidence interval will be between $\hat{p} + z_{1-\frac{\alpha}{2} } \sqrt{ \frac{1}{n} \hat{p} (1-\hat {p})} \text{ and } \hat{p} - z_{1-\frac{\alpha}{2} } \sqrt{ \frac{1}{n} \hat{p} (1-\hat {p})} . $ By substituting into these expressions, we find that the confidence interval is between \begin{align} &\dfrac{500}{1050} + 1.645 \sqrt{ \frac{1}{1050} \times \dfrac{500}{1050} \times \left(1- \dfrac{500}{1050}\right) }\\ \text{ and } &\dfrac{500}{1050} - 1.645 \sqrt{ \frac{1}{1050} \times \dfrac{500}{1050} \times \left(1- \dfrac{500}{1050}\right) }\\\\ &=0.47619 + (1.645 \times \sqrt{0.00024} ) \text{ and } 0.47619 - (1.645 \times \sqrt{0.00024} ) \\ &=0.50155 \text{ and } 0.45084 . \end{align} So the confidence interval is $(0.45084, 0.50155)$. Normal DistributionWe can use either the $z$-score or the sample mean $\bar{x}$ as the test statistic. If the $z$-score is used then reading straight from the tables gives the critical values. For example, the critical values for a $5$% significance test are: To obtain a confidence interval for the mean, use the following procedure: For a two-tailed test with a $5$% significance level we need to consider \begin{align} 0.95 &= \mathrm{P}[-k< Z < k] \\ &= \mathrm{P}\left[-k<\dfrac{\bar{X}-\mu}{\frac{\sigma}{\sqrt{n} } } \mu+1.96\frac{\sigma}{\sqrt{n} }. \] Student $t$-distributionGiven the number of degrees of freedom $v$ and the significance level $\alpha$, the critical values can be obtained from the tables. Critical regions can then be computed from these. If we are performing a hypothesis test at a $1$% significance level with $15$ degrees of freedom using the Student $t$-distribution then there are three cases, depending on the alternative hypothesis. If we are performing a two-tailed test, the critical values are $\pm2.9467$ so the confidence interval is $-2.9467 \leq t \leq 2.9467$ where $t$ is the test statistic. The critical regions will be $t< -2.9467$ and $t>2.9467$. If we are performing a one-tailed test, the critical value is $2.6025$: Video ExamplesIn this video, Daniel Organisciak calculates a one-tailed confidence interval for the normal distribution. In this video Daniel Organisciak calculates a two-tailed confidence interval for the normal distribution. Confidence intervals and hypothesis testing- Understand the t value and Pr(>|t|) fields in the output of lm
- Be able to think critically about the meaning and limitations of strict hypothesis tests
Confidence intervals and hypothesis testsT-statistics. Suppose we’re interested in the value \(\beta_k\) , the \(k\) –th entry of \(\betav\) in for some regression \(\y_n \sim \betav^\trans \xv_n\) . Recall that we have been finding \(\v\) such that \[ \sqrt{N} (\beta_k - \beta) \rightarrow \gauss{0, \v}. \] For example, under homoskedastic assumptions with \(\y_n = \xv_n^\trans \beta + \res_n\) , we have \[ \begin{aligned} \v =& \sigma^2 (\Xcov^{-1})_{kk} \textrm{ where } \\ \Xcov =& \lim_{N \rightarrow \infty} \frac{1}{N} \X^\trans \X \textrm{ and } \\ \sigma^2 =& \var{\res_n}. \end{aligned} \] Typically we don’t know \(\v\) , but have \(\hat\v\) such that \(\hat\v \rightarrow \v\) as \(N \rightarrow \infty\) . Again, under homoeskedastic assumptions, \[ \begin{aligned} \hat\v =& \hat\sigma^2 \left(\frac{1}{N} \X^\trans \X \right)_{kk} \textrm{ where } \\ \hat\sigma^2 =& \frac{1}{N-P} \sumn \reshat_n^2. \end{aligned} \] Putting all this together, the quantity \[ \t = \frac{\sqrt{N} (\betahat_k - \beta_k)}{\sqrt{\hat\v}} = \frac{\betahat_k - \beta_k}{\sqrt{\hat\v / N}} \] has an approximately standard normal distribution for large \(N\) . Quantities of this form are called “T–statistics,” since, under our normal assumptions, we have shown that \[ \t \sim \studentt{N-P}, \] exactly for all \(N\) . Despite it’s name, it’s worth remembering that a T–statistic is actually not Student T distributed in general; it is asymptotically normal. Recall that for large \(N\) , the Student T and standard normal distributions coincide. Plugging in values for \(\beta_k\)However, there’s something funny about a “T-statistic” — as written, you cannot compute it, because you don’t know \(\beta_k\) . In fact, finding what values \(\beta_k\) might plausibly take is the whole point of statistical inference. So what good is a T–statistic? Informally, one way to reason about it is as follows. Let’s take some concrete values for an example. Suppose guess that \(\beta_k^0\) is the value, and compute \[ \betahat_k = 2 \quad\textrm{and}\quad \sqrt{\hat\v / N} = 3 \quad\textrm{so}\quad \t = \frac{2 - \beta_k^0}{3}. \] We use the superscript \(0\) to indicate that \(\beta_k^0\) is our guess, not necessarily the true value. Suppose we plug in some particular value, such as \(\beta_k^0 = 32\) . Using this value, we compute our T–statistic, and find that it’s very large — in our example, we would have \(\t = (2 - 32) / 3 = -30\) . It’s very unlikely to get a standard normal (or Student T) draw this large. Therefore, either: - We got a very (very very very very) unusual draw of our standard normal or
- We guessed wrong, i.e. \(\beta_k \ne \beta_k^0 = 32\) .
In this way, we might consider it plausible to “reject” the hypothesis that \(\beta_k = 32\) . There’s a subtle problem with the preceding reasoning, however. Suppose we do the same calculation with \(\beta_k^0 = 1\) . Then \(\t = (2 - 1) / 3 = 1/3\) . This is a much more typical value for a standard normal distribution. However, the probability of getting exactly \(1/3\) — or, indeed, any particular value — is zero, since the normal distribution is continuous valued. (This problem is easiest to see with continuous random variables, but the same basic problem will occur when the distribution is discrete but spread over a large number of possible values.) Rejection regionsTo resolve this problem, we can specify regions that we consider implausible. That is, suppose we take a region \(R\) such that, if \(\t\) is standard normal (or Student-T), then \[ \prob{\t \in R} \le \alpha \quad\textrm{form some small }\alpha. \] For example, we might take \(\Phi^{-1}(\cdot)\) to be the inverse CDF of \(\t\) if \(\beta_k = \beta_k^0\) . Then we can take \[ R_{ts} = \{\t: \abs{t} \ge q \} \quad\textrm{where } q = \Phi^{-1}(\alpha / 2)\\ \] where \(q\) is an \(\alpha / 2\) quantile of the distribution of \(\t\) . But there are other choices, such as \[ \begin{aligned} R_{u} ={}& \{\t: \t \ge q \} \quad\textrm{where } q = \Phi^{-1}(1 - \alpha) \\ R_{l} ={}& \{\t: \t \le q \} \quad\textrm{where } q = \Phi^{-1}(\alpha) \\ R_{m} ={}& \{\t: \abs{\t} \le q \} \quad\textrm{where } q = \Phi^{-1}(0.5 + \alpha / 2) \quad\textrm{(!!!)}\\ R_{\infty} ={}& \begin{cases} \emptyset & \textrm{ with independent probability } \alpha \\ (-\infty,\infty) & \textrm{ with independent probability } 1 - \alpha \\ \end{cases} \quad\textrm{(!!!)} \end{aligned} \] The last two may seem silly, but they are still rejection regions into which \(\t\) is unlikely to fall if it has a standard normal distribution. How can we think about \(\alpha\) , and about the choice of the region? Recall that - If \(\t \in R\) , we “reject” the proposed value of \(\beta_k^0\)
- If \(\t \notin R\) , we “fail to reject” the given value of \(\beta_k^0\) .
Of course, we don’t “accept” the value of \(\beta_k^0\) in the sense of believing that \(\beta_k^0 = \beta_k\) — if nothing else, there will always be multiple values of \(\beta_k^0\) that we do not reject, and \(\beta_k\) cannot be equal to all of them. So there are two ways to make an error: - Type I error: We are correct and \(\beta_k = \beta_k^0\) , but \(\t \in R\) and we reject
- Type II error: We are incorrect and \(\beta_k \ne \beta_k^0\) , but \(\t \notin R\) and we fail to reject
By definition of the region \(R\) , we have that \[ \prob{\textrm{Type I error}} \le \alpha. \] This is true for all the regions above, including the silly ones! What about the Type II error? It must depend on the “true” value of \(\beta_k\) , and on the shape of the rejection region we choose. Note that \[ \t = \frac{\betahat_k - \beta_k^0}{\sqrt{\hat\v / N}} = \frac{\betahat_k - \beta_k}{\sqrt{\hat\v / N}} + \frac{\beta_k - \beta_k^0}{\sqrt{\hat\v / N}} \] So if the true value \(\beta_k \gg \beta_k^0\) , then our \(\t\) statistic is too large, and so on. For example: - Then \(\t\) is too large and positive.
- \(R_u\) and \(R_{ts}\) will reject, but \(R_l\) will not.
- The Type II error of \(R_u\) will be lowest, then \(R_{ts}\) , then \(R_l\) .
- \(R_l\) actually has greater Type II error than the silly regions, \(R_\infty\) and \(R_m\) .
- Then \(\t\) is too large and negative.
- \(R_l\) and \(R_{ts}\) will reject, but \(R_u\) will not.
- The Type II error of \(R_l\) will be lowest, then \(R_{ts}\) , then \(R_u\) .
- \(R_u\) actually has greater Type II error than the silly regions, \(R_\infty\) and \(R_m\) .
- Then \(\t\) has about the same distribution as when \(\beta_k^0 = \beta_k\) .
- All the regions reject just about as often as we commit a Type I error, that is, a proportion \(\alpha\) of the time.
Thus the shape of the region determines which alternatives you are able to reject. The probability of “rejecting” under a particular alternative is called the “power” of a test; the power is one minus the Type II error rate. The null and alternativeStatistics has some formal language to distinguish between the “guess” \(\beta_k^0\) and other values. - Falsely rejecting the null hypothesis is called a Type I error
- By construction, Type I errors occurs with probability at most \(\alpha\)
- Falsely failling to reject the null hypothesis is called a Type II error
- Type II errors’ probability depends on the alternative(s) and the rejection region shape.
The choice of a test statistic (here, \(\t\) ), together with a rejection region (here, \(R\) ) constitute a “test” of the null hypothesis. In general, one can imagine constructing many different tests, with different theoretical guarantees and power. Confidence intervalsOften in applied statistics, a big deal is made about a single hypothesis test, particularly the null that \(\beta_k^0 = 0\) . Often this is not a good idea. Typically, we do not care whether \(\beta_k\) is precisely zero; rather, we care about the set of plausible values \(\beta_k\) might take. The distinction can be expressed as the difference between statistical and practical significance: - Statistical significance is the size of an effect relative to sampling variability
- Practical significance is the size of the effect in terms of its effect on reality.
For example, suppose that \(\beta_k\) is nonzero but very small, but \(\sqrt{\hat\v / N}\) is very small, too. We might reject the null hypothesis \(\beta_k^0 = 0\) with a high degree of certainty, and call our result statistically significant . However, a small value of \(\beta_k\) may still not be a meaningful effect size for the problem at hand, i.e., it may not be practically significant . A remendy is confidence intervals, which are actually closely related to our hypothesis tests. Recall that we have been constructing intervals of the form \[ \prob{\beta_k \in I} \ge 1-\alpha \] \[ I = \left(\betahat_k \pm q \hat\v / \sqrt{N}\right), \] where \(q = \Phi^{-1}(\alpha / 2)\) , and \(\Phi\) is the CDF of either the standard normal or Student T distribution. It turns out that \(I\) is precisely the set of values that we would not reject with region \(R_{ts}\) . And, indeed, given a confidence interval, a valid test of the hypothesis \(\beta_k^0\) is given by rejecting if an only if \(\beta_k^0 \in I\) . This duality is entirely general: - The set of values that a valid test does not reject is a valid confidence interval
- Checking whether a value falls in a valid confidence interval is a valid test
Source CodeData Science Tutorials For Data Science Learners - Select variables of data frame in R R
- Sort or Order Rank in R R
- Locate position of patterns in a character string in R R
Rejection Region in Hypothesis TestingRejection Region in Hypothesis Testing, What is the definition of a Rejection Region? A rejection region is a section of a graph where the null hypothesis is rejected (assuming your test results fall into that area). Rejection RegionThe primary goal of statistics is to test theories or experiment results. For example, you may have developed a novel fertilizer that you believe accelerates plant growth by 50%. To demonstrate that your hypothesis is correct, your experiment must: Be consistent. Be likened to a well-known plant fact (in this example, probably the average growth rate of plants without the fertilizer). This sort of statistical testing is known as a hypothesis test. Descriptive statistics vs Inferential statistics: Guide The testing process includes a rejection region (also known as a crucial region). It is a branch of probability that determines if your theory (or “hypothesis”) is likely to be correct. Probability Distributions and Rejection Regions rejection region A two-tailed t-rejection distribution’s zones. A probability distribution can be used to draw every rejection region. A two-tailed t-distribution is seen in the figure above. A rejection region can also be seen in just one tail. Two-Tailed vs One-TailedYour null hypothesis statement determines the type of test to use. If your question is, “Is the average growth rate larger than 8cm per day?”. Because you’re only interested in one way, this is a one-tailed test (greater than 8cm a day). You might alternatively have a single “less than” rejection region. “Is the growth rate less than 8cm per day?” for example. When you want to see if there’s a difference in both directions, you’ll utilize a two-tailed test with two regions (greater than and less than). Alpha Levels and Rejection RegionsAs a researcher, you decide what amount of alpha you’re willing to take. For example, if you wanted to be 95% confident that your results are significant, you would set a 5% alpha level (100% – 95%). That 5% threshold is the rejection threshold. In a one-tailed test, the 5% would be in one of the tails. The rejection zone for a two-tailed test would be in two tails. The rejection zone is in one tail of a one-tailed test. P-Values and Rejection RegionsA hypothesis can be tested in two ways: with a p-value or with a critical value. p-value method: When you perform a hypothesis test (for example, a z test), the result is a p-value. A “probability value” is the p-value. It’s what determines if your hypothesis is likely to be correct or not. If the value falls within the rejection range, the results are statistically significant, and the null hypothesis can be rejected. If your p-value is beyond the rejection range, your results are insufficient to reject the null hypothesis. What is statistical significance? A statistically significant outcome in the case of plant fertilizer would be one that shows the fertilizer does actually make plants grow quicker (compared to other fertilizers). The stages are the same for the Rejection Region approach with a critical value. Instead of calculating a p-value, a crucial value is calculated. If the value is within the region, the null hypothesis is rejected. Related PostsLeave a Reply Cancel replyYour email address will not be published. Required fields are marked * Save my name, email, and website in this browser for the next time I comment. Yes, add me to your mailing list User PreferencesContent preview. Arcu felis bibendum ut tristique et egestas quis: - Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris
- Duis aute irure dolor in reprehenderit in voluptate
- Excepteur sint occaecat cupidatat non proident
Keyboard Shortcuts6a.4.3 - steps in conducting a hypothesis test for \(p\), six steps for one-sample proportion hypothesis test, steps 1-3 section . Let's apply the general steps for hypothesis testing to the specific case of testing a one-sample proportion. \( np_0\ge 5 \) and \(n(1−p_0)≥5 \) One Proportion Z-test Hypotheses One Proportion Z-test: \(z^*=\dfrac{\hat{p}-p_0}{\sqrt{\frac{p_0(1-p_0)}{n}}} \) Rejection Region ApproachSteps 4-6 section , left-tailed test, right-tailed test, two-tailed test. Reject \(H_0\) if \(z^* \le z_\alpha\) Reject \(H_0\) if \(z^* \ge z_{1-\alpha}\) Reject \(H_0\) if \(|z^*| \ge |z_{\alpha/2}|\) View the critical values and regions with an \(\alpha=.05\). These graphs show the various z-critical values for tests at an \(\alpha=.05\). *The graphs are not to scale. Reject \(H_0\) if \(z^* \le -1.65\) Reject \(H_0\) if \(z^* \ge 1.65\) Reject \(H_0\) if \(|z^*| \ge |-1.96|\) P-Value ApproachExample 6-5: penn state students from pennsylvania section . Referring back to example 6-4. Say we take a random sample of 500 Penn State students and find that 278 are from Pennsylvania. Can we conclude that the proportion is larger than 0.5 at a 5% level of significance? Conduct the test using both the rejection region and p-value approach. - Steps 4-6: Rejection Region
- Steps 4-6: P-Value
Set up the hypotheses. Since the research hypothesis is to check whether the proportion is greater than 0.5 we set it up as a one (right)-tailed test: \( H_0\colon p=0.5 \) vs \(H_a\colon p>0.5 \) Can we use the z-test statistic? The answer is yes since the hypothesized value \(p_0 \) is \(0.5\) and we can check that: \(np_0=500(0.5)=250 \ge 5 \) and \(n(1-p_0)=500(1-0.5)=250 \ge 5 \) According to the question, \(\alpha= 0.05 \). \begin{align} z^*&= \dfrac{0.556-0.5}{\sqrt{\frac{0.5(1-0.5)}{500}}}\\z^*&=2.504 \end{align} We can use the standard normal table to find the value of \(Z_{0.05} \). From the table, \(Z_{0.05} \) is found to be \(1.645\) and thus the critical value is \(1.645\). The rejection region for the right-tailed test is given by: \( z^*>1.645 \) The test statistic or the observed Z-value is \(2.504\). Since \(z^*\) falls within the rejection region, we reject \(H_0 \). With a test statistic of \(2.504\) and critical value of \(1.645\) at a 5% level of significance, we have enough statistical evidence to reject the null hypothesis. We conclude that a majority of the students are from Pennsylvania. Since \(\text{p-value} = 0.0062 \le 0.05\) (the \(\alpha \) value), we reject the null hypothesis. With a test statistic of \(2.504\) and p-value of \(0.0062\), we reject the null hypothesis at a 5% level of significance. We conclude that a majority of the students are from Pennsylvania. Online Purchases Section An e-commerce research company claims that 60% or more graduate students have bought merchandise online. A consumer group is suspicious of the claim and thinks that the proportion is lower than 60%. A random sample of 80 graduate students shows that only 22 students have ever done so. Is there enough evidence to show that the true proportion is lower than 60%? Conduct the test at 10% Type I error rate and use the p-value and rejection region approaches. Set up the hypotheses. Since the research hypothesis is to check whether the proportion is less than 0.6 we set it up as a one (left)-tailed test: \( H_0\colon p=0.6 \) vs \(H_a\colon p<0.6 \) Can we use the z-test statistic? The answer is yes since the hypothesized value \(p_0 \) is 0.6 and we can check that: \(np_0=80(0.6)=48 \ge 5 \) and \(n(1-p_0)=80(1-0.6)=32 \ge 5 \) According to the question, \(\alpha= 0.1 \). \begin{align} z^* &=\frac{\hat{p}-p_0}{\sqrt{\frac{p_0(1-p_0)}{n}}}\\&=\frac{.275-0.6}{\sqrt{\frac{0.6(1-0.6)}{80}}}\\&=-5.93 \end{align} The critical value is the value of the standard normal where 10% fall below it. Using the standard normal table, we can see that the value is -1.28. The rejection region is any \(z^* \) such that \(z^*<-1.28 \) . Since our test statistic, -5.93, is inside the rejection region, we reject the null hypothesis. There is enough evidence in the data provided to suggest, at 10% level of significance, that the true proportion of students who made purchases online was less than 60%. Since our p-value is very small and less than our significance level of 10%, we reject the null hypothesis. Stack Exchange NetworkStack Exchange network consists of 183 Q&A communities including Stack Overflow , the largest, most trusted online community for developers to learn, share their knowledge, and build their careers. Q&A for work Connect and share knowledge within a single location that is structured and easy to search. Confused about rejection region and P-valueI am confused about the rejection region and P-value. I thought that the P-value is the simply the probability associated to the set of points where we would reject the null hypothesis (the rejection region). But according to this response, they are not related. Is it then possible, having $R$ as a test statstic, to have a rejection region, say $C=\{R^{observed}:R^{observed}\geq w_{1−\frac{\alpha}{2}}\}\cup\{R^{observed}:R^{observed}\leq w_{\frac{\alpha}{2}}\}$ , but the p-value is $P(R>R^{observed})$ ? 2 Answers 2I think this will be best understood with an example. Let us solve a hypothesis test for the mean height of people in a country. We have the information about the heights of a sample of people in that country. First, we define our null and alternative hypthesis: - $H_0: \mu \geq a$
- $H_1: \mu < a$
And (let me change your notation) we have our test statistic $Z$ . Now, we know two things about this test statistic: We know the formula for this statistic: $Z=\dfrac{\bar{x}-\mu}{\sigma/\sqrt{n}}$ We know the distribution it follows. For simplicity in this explanation, let me assume it follows a normal distribution. We would have that $Z\sim N(0,1)$ . Now the important part: We do not know the true population value of $\mu$ (this means, we do not know the true mean height of all the people in the country, if we wanted to know that we would need to know the height of all the citizens in that country). But we can say: hey, let's assume that the value for $\mu$ is the value stated in $H_0$ (a.k.a. let's assume that $\mu=a$ ) And we pose the key question: How likely is it for $\mu$ to take the value $a$ given the information that we know from the sample data? Now that we are assuming that $\mu=a$ , we can obtain the value of our test statistic under the null hypothesis (this is, assuming that $\mu=a$ ). There can be two possible results: - If $a$ is not a likely value for $\mu$ to have then the statistic $\hat{Z}$ value will not fit well in the distribution $Z$ follows and we will reject $H_0$
- If $a$ is a likely value for $\mu$ to have, then the statistic $\hat{Z}$ value will fit well in the distribution $Z$ follows and we will fail to reject $H_0$ :
And finally here comes into play the rejection region and the p-value. - We will consider that the tail of the distribution (in this case, the left tail, as stated by $H_1$ are all not likely values for $\hat{Z}$ so if any value is close to the tail, we reject $H_0$ . How close to the tail? That is stated by the significance level $\alpha$ . The rejection region is:
$$RR=\{Z {\ }s.t.{\ } Z < -Z_{\alpha}\}$$ If we take, for example, $\alpha=0.05$ then the rejection region is $$RR=\{Z {\ }s.t.{\ } Z < -Z_{0.05}\}= \{Z {\ }s.t.{\ } Z < -1.645\}$$ - And the p-value is simply the probability of obtaining a value at least as extreme as the one from our sample, or in other words, if the sample value of our statistic is $\hat{Z}$ then the p-value is $$p-value=P(Z<\hat{Z})$$
In one image, in red the rejection region, and in green the p-value. Remark : This plots have been made assuming that we are doing a left sided test . Considering a right sided or two sided test would yield similar but not equal images. - $\begingroup$ Thank you for your explanation! So If we wanted to test $H_0: \mu =a$ vs $H_1:\mu \neq a$, then $RR=\{Z s.t. Z_{1-\frac{\alpha}{2}} <Z < Z_{\frac{\alpha}{2}}\}$, and the P-value would be the probability $P(|Z|<\hat{Z})$. Right? $\endgroup$ – Toney Shields Commented Feb 9, 2021 at 10:35
- 1 $\begingroup$ Yeah, that is right. $\endgroup$ – Álvaro Méndez Civieta Commented Feb 9, 2021 at 10:52
- $\begingroup$ Suppose then we want to test some other parameter $H_0: \theta = a$ vs $H_1: \theta \neq a$, and that our test statistic $T$ under the null hypothesis has a Gamma distribution $\Gamma(m,n)$. The rejection region would still the same except that the quantile changes to that of a $\Gamma(m,n)$. What I'm having trouble with is : what is the P-value in this case? $\endgroup$ – Toney Shields Commented Feb 9, 2021 at 10:58
- 1 $\begingroup$ Well in two sided tests as you said you can always obtain the rejection region, but I am afraid that the two sided p-value is only well defined when the test statistic has a symetric distribution. $\endgroup$ – Álvaro Méndez Civieta Commented Feb 9, 2021 at 11:08
The rejection region is fixed beforehand. If the null hypothesis is true then some $\alpha \%$ of the observations will be in the region. The p-value is not the same as this $\alpha \%$ . The p-value is computed for each separate observation, and can be different for two observations that both fall inside the rejection region. The p-value indicates how extreme* a value is. And expresses this in terms of a probability. This expression in terms of a probability could be seen as the quantile of the outcome when the potential outcomes are ranked in decreasing order of extremity. The more extreme the observation, the lower the quantile. In short: The rejection region can be seen as the region of observations for which the associated quantile or p-value is lower than some value. See also: https://stats.stackexchange.com/questions/tagged/critical-value * What is and what is not considered extreme is not well defined here and might be considered arbitrary, but depending on the situation there might be good reasons to choose a particular definition. For example, think about one-sided and two-sided tests in which case different sorts of extremities are chosen. Because of the variations in choice for 'extremeness', it might be that you encounter a situation where some observation is inside the rejection region but has a p-value that is larger. This is the case when the two use a different definition. But typically the p-value and rejection region should relate to the same definition of 'extremeness'. - $\begingroup$ Does it mean that the rejection region can be $C=\{R^{observed}:R^{observed}\geq w_{1−\frac{\alpha}{2}}\}\cup\{R^{observed}:R^{observed}\leq w_{\frac{\alpha}{2}}\}$, but a P-value $P(R>R^{observed})$ (meaning that "extreme" is when $R>R^{observed}$)? $\endgroup$ – Toney Shields Commented Feb 9, 2021 at 11:02
- 1 $\begingroup$ Ah now I see your problem. $P(R>R^{observed})$ can be very high. Say you have the hypothesis $R \sim N(0,\sigma^2)$ (ie normal distributed). The 5% rejection region could be when for the absolute value $|R|>2\sigma$. In that case, if you have an observation below $-2$, that is $R^{observed}<-2$, then the observation is inside the rejection region, but the probability to observe $R>R^{observed}$ is very high... $\endgroup$ – Sextus Empiricus Commented Feb 9, 2021 at 11:15
- 1 $\begingroup$ ... the discrepancy occurs because the probability $P(R>R^{observed})$ is not using the same definition for an extreme value as the definition that has been used for the rejection region. $\endgroup$ – Sextus Empiricus Commented Feb 9, 2021 at 11:16
- 1 $\begingroup$ @ToneyShields, this has to do with the original Fisherian understanding of $p$-value (how extreme the result is as determined by $H_0$ alone, i.e. the sampling distribution of the test statistic under $H_0$) versus the modern Fisher-Neyman-Pearson hybrid ((how extreme the result is as determined by $H_0$ and $H_1$ together). I have a few related threads here . In that (and other) regard(s), the footnote of Sextus' answer is an important one. $\endgroup$ – Richard Hardy Commented Feb 9, 2021 at 12:41
Your AnswerSign up or log in, post as a guest. Required, but never shown By clicking “Post Your Answer”, you agree to our terms of service and acknowledge you have read our privacy policy . Not the answer you're looking for? Browse other questions tagged hypothesis-testing p-value or ask your own question .- Featured on Meta
- Preventing unauthorized automated access to the network
- User activation: Learnings and opportunities
- Join Stack Overflow’s CEO and me for the first Stack IRL Community Event in...
Hot Network Questions- Does this work for page turns in a busy violin part?
- Why would elves care what happens to Middle Earth?
- removing dead citrix apps from applications menu?
- Macro not working in newcommand, but plain text input works
- Is it possible to know where the Sun is just by looking at the Moon?
- Do early termination fees hold up in court?
- What are major reasons why Republicans support the death penalty?
- Are logic and mathematics the only fields in which certainty (proof) can be obtained?
- How to fix bottom of stainless steel pot that has been separated from its main body?
- Undamaged tire repeatedly deflating
- CH in non-set theoretic foundations
- Is it possible to speed up this function?
- Are there individual protons and neutrons in a nucleus?
- How do you tell someone to offer something to everyone in a room by taking it physically to everyone in the room so everyone can have it?
- How can I make second argument path relative to the first on a command?
- Tikz: On straight lines moving balls on a circle inside a regular polygon
- What does IpOutDiscard in RFC1213 / nstat actually mean?
- Is there any language which distinguishes between “un” as in “not” and “un” as in “inverse”?
- Does legislation on transgender healthcare affect medical researchers?
- If two subgroups intersect in only the identity, do their cosets intersect in at most one element?
- Is “No Time To Die” the first Bond film to feature children?
- FIFO capture using cat not working as intended?
- Neil Tyson: gravity is the same every where on the geoid
- How to sub-align expressions and preserve equation numbering?
|
IMAGES
VIDEO
COMMENTS
In this situation, the rejection region is on the right side. So, if the test statistic is bigger than the cut-off z-score, we would reject the null, otherwise, we wouldn't. Importance of the Significance Level and the Rejection Region. To sum up, the significance level and the reject region are quite crucial in the process of hypothesis ...
A one tailed test with the rejection region in one tail. Rejection Regions and P-Values. There are two ways you can test a hypothesis: with a p-value and with a critical value. P-value method: When you run a hypothesis test (for example, a z test), the result of that test will be a p value. The p value is a "probability value."
The critical value for conducting the left-tailed test H0 : μ = 3 versus HA : μ < 3 is the t -value, denoted -t(α, n - 1), such that the probability to the left of it is α. It can be shown using either statistical software or a t -table that the critical value -t0.05,14 is -1.7613. That is, we would reject the null hypothesis H0 : μ = 3 in ...
Make a decision about the null hypothesis: In this step, we decide to either reject the null hypothesis or decide to fail to reject the null hypothesis. Notice we do not make a decision where we will accept the null hypothesis. State an overall conclusion: Once we have found the p-value or rejection region, and made a statistical decision about ...
The rejection region is the region where, if our test statistic falls, then we have enough evidence to reject the null hypothesis. If we consider the right-tailed test, for example, the rejection region is any value greater than \(c_{1-\alpha} \), where \(c_{1-\alpha}\) is the critical value.
Figure \(\PageIndex{1}\): The rejection region for a one-tailed test. The shaded rejection region takes us 5% of the area under the curve. Any result which falls in that region is sufficient evidence to reject the null hypothesis. The rejection region is bounded by a specific \(z\)-value, as is any area under the curve.
The rejection region is bounded by a specific z value, as is any area under the curve. In hypothesis testing, the value corresponding to a specific rejection region is called the critical value, z crit (" z crit"), or z * (hence the other name "critical region"). Finding the critical value works exactly the same as finding the z score corresponding to any area under the curve as we did ...
Every instance of hypothesis testing discussed in this and the following two chapters will have a rejection region like one of the six forms tabulated in the tables above. No matter what the context a test of hypotheses can always be performed by applying the following systematic procedure, which will be illustrated in the examples in the ...
In the context of the marketing team's hypothesis testing, the reject region for the one-tailed test with an alpha level of 1% corresponds to the range of z-scores that fall within the top 1% of the normal distribution. ... It also circles the rejection regions of null hypothesis when z = 2.33 and alpha = 0.01.
A statistical hypothesis test is a method of statistical inference used to decide whether the data sufficiently supports a ... These define a rejection region for each hypothesis. 2 Report the exact level of significance (e.g. p = 0.051 or p = 0.049). Do not refer to "accepting" or "rejecting" hypotheses. If the result is "not significant ...
If the null hypothesis is false, then the F statistic will be large. The rejection region for the F test is always in the upper (right-hand) tail of the distribution as shown below. Rejection Region for F Test with a =0.05, df 1 =3 and df 2 =36 (k=4, N=40) For the scenario depicted here, the decision rule is: Reject H 0 if F > 2.87. The ANOVA ...
Two-sided hypothesis tests have two rejection regions. Consequently, you'll need two critical values that define them. Because there are two rejection regions, we must split our significance level in half. Each rejection region has a probability of α / 2, making the total likelihood for both areas equal the significance level.
This region, which leads to rejection of the null hypothesis, is called the rejection region. For example, for a significance level of 5%: For an upper-tail test, the critical value is the 95th percentile of the t-distribution with n−1 degrees of freedom; reject the null in favor of the alternative if the t-statistic is greater than this.
The procedure for hypothesis testing is based on the ideas described above. Specifically, we set up competing hypotheses, select a random sample from the population of interest and compute summary statistics. We then determine whether the sample data supports the null or alternative hypotheses. ... Rejection Region for Two-Tailed Z Test (H 1 ...
A critical region, also known as the rejection region, is a set of values for the test statistic for which the null hypothesis is rejected. i.e. if the observed test statistic is in the critical region then we reject the null hypothesis and accept the alternative hypothesis. Critical Values
In this approach, we construct the rejection region under the probability density curve of the involved distribution based on the type of the test and the significance level alpha. The boundaries of the region are called the critical values and can be found in a similar way for all procedures. ... Due to the logic of a hypothesis testing ...
Type II errors' probability depends on the alternative(s) and the rejection region shape. The choice of a test statistic (here, \(\t\)), together with a rejection region (here, \(R\)) constitute a "test" of the null hypothesis. In general, one can imagine constructing many different tests, with different theoretical guarantees and power.
The intent of hypothesis testing is formally examine two opposing conjectures (hypotheses), H0 and HA. These two hypotheses are mutually exclusive and exhaustive so that one is true to the exclusion of the other. We accumulate evidence - collect and analyze sample information - for the purpose of determining which of the two hypotheses is true ...
The significance level is the probability of getting a result in the rejection region, given the null hypothesis is true. Note that the alternative puts an ordering on your test statistic - the values of the test statistic most in keeping with the alternative are the ones you want in your rejection region.. The p-value is the probability of a test statistic at least as extreme (under that ...
Using the rejection region approach, you need to check the table or software for the critical value every time you use a different α value. In addition to just using it to reject or not reject H 0 by comparing p-value to α value, the p-value also gives us some idea of the strength of the evidence against H 0.
The testing process includes a rejection region (also known as a crucial region). It is a branch of probability that determines if your theory (or "hypothesis") is likely to be correct. Probability Distributions and Rejection Regions rejection region. A two-tailed t-rejection distribution's zones. A probability distribution can be used to ...
You could even choose a region such as $[\mu-d,\mu + d]$ ! It is however not natural, because one would like your rejection set to include weird values of the statistic (values that are far from $\mu$) rather than normal values. Why? Because the test statistic is expected to be a measure of the distance between the data and the null hypothesis ...
Write down clearly the rejection region for the problem. The critical value is the value of the standard normal where 10% fall below it. Using the standard normal table, we can see that the value is -1.28. Step 5: Make a decision about the null hypothesis. The rejection region is any \(z^* \) such that \(z^*<-1.28 \) .
The rejection region is fixed beforehand. If the null hypothesis is true then some α% of the observations will be in the region. The p-value is not the same as this α%. The p-value is computed for each separate observation, and can be different for two observations that both fall inside the rejection region. The p-value indicates how extreme ...
The Null hypothesis \(\left(H_{O}\right)\) is a statement about the comparisons, e.g., between a sample statistic and the population, or between two treatment groups. The former is referred to as a one-tailed test whereas the latter is called a two-tailed test. The null hypothesis is typically "no statistical difference" between the ...