alternative assumption to the null hypothesis
Null hypothesis and Alternative Hypothesis
Why do we need alternative hypothesis? - Cross Validated
Should the alternative hypothesis always be the research hypothesis? - Cross Validated
The alternative hypothesis
Conditional on the data, you cannot confirm the alternative hypothesis but you can reject the null. It sounds to me that what you are being asked is misleading because it depends on the context.
More on reddit.comWhatโs the difference between a research hypothesis and a statistical hypothesis?
What are null and alternative hypotheses?
What is hypothesis testing?
Videos
Hey! Can someone explain to me in simple terms the definition of null hypothesis? If u can use an example it would be great! Also if we reject the null hypothesis does it mean that the alternative hypothesis is true?
There was, historically, disagreement about whether an alternative hypothesis was necessary. Let me explain this point of disagreement by considering the opinions of Fisher and Neyman, within the context of frequentist statistics, and a Bayesian answer.
Fisher - We do not need an alternative hypothesis; we can simply test a null hypothesis using a goodness-of-fit test. The outcome is a
-value, providing a measure of evidence for the null hypothesis.
Neyman - We must perform a hypothesis test between a null and an alternative. The test is such that it would result in type-1 errors at a fixed, pre-specified rate,
. The outcome is a decision - to reject or not reject the null hypothesis at the level
.
We need an alternative from a decision theoretic perspective - we are making a choice between two courses of action - and because we should report the power of the test $$ 1 - p\left(\textrm{Accept $H_0$} \, \middle|\, H_1\right) $$ We should seek the most powerful tests possible to have the best chance of rejecting
when the alternative is true.
To satisfy both these points, the alternative hypothesis cannot be the vague 'not
' one.
Bayesian - We must consider at least two models and update their relative plausibility with data. With only a single model, we simple have
no matter what data we collect. To make calculations in this framework, the alternative hypothesis (or model as it would be known in this context) cannot be the ill-defined 'not
' one. I call it ill-defined since we cannot write the model
.
I will focus on "If we do not talk about accepting alternative hypothesis, why do we need to have alternative hypothesis at all?"
Because it helps us to choose a meaningful test statistic and design our study to have high power---a high chance of rejecting the null when the alternative is true. Without an alternative, we have no concept of power.
Imagine we only have a null hypothesis and no alternative. Then there's no guidance on how to choose a test statistic that will have high power. All we can say is, "Reject the null whenever you observe a test statistic whose value is unlikely under the null." We can pick something arbitrary: we could draw Uniform(0,1) random numbers and reject the null when they are below 0.05. This happens under the null "rarely," no more than 5% of the time---yet it's also just as rare when the null is false. So this is technically a statistical test, but it's meaningless as evidence for or against anything.
Instead, usually we have some scientifically-plausible alternative hypothesis ("There is a positive difference in outcomes between the treatment and control groups in my experiment"). We'd like to defend it against potential critics who would bring up the null hypothesis as devil's advocates ("I'm not convinced yet---maybe your treatment actually hurts, or has no effect at all, and any apparent difference in the data is due only to sampling variation").
With these 2 hypotheses in mind, now we can setup up a powerful test, by choosing a test statistic whose typical values under the alternative are unlikely under the null. (A positive 2-sample t-statistic far from 0 would be unsurprising if the alternative is true, but surprising if the null is true.) Then we figure out the test statistic's sampling distribution under the null, so we can calculate p-values---and interpret them. When we observe a test statistic that's unlikely under the null, especially if the study design, sample size, etc. were chosen to have high power, this provides some evidence for the alternative.
So, why don't we talk about "accepting" the alternative hypothesis? Because even a high-powered study doesn't provide completely rigorous proof that the null is wrong. It's still a kind of evidence, but weaker than some other kinds of evidence.
I would say that the "alternative hypothesis" is usually NOT a "proposed hypothesis".
You do not define "proposed hypothesis" and it is not a common phrase. Presumably you mean that it is either a statistical hypothesis or it is a scientific hypothesis. They are usually quite different things.
A scientific hypothesis usually concerns a something to do with the true state of the real world, whereas a statistical hypothesis concerns only conditions within a statistical model. It is very common for the real world to be more complicated and less well-defined than a statistical model and so inferences regarding a statistical hypothesis will need to be thoughtfully extrapolated to become relevant to a scientific hypothesis.
For your example a scientific hypothesis concerning the two drugs in question might be something like 'drug x can be substituted for drug y without any noticeable change in results experienced by the patients'. A relevant statistical hypothesis would be much more restricted along the lines of 'drug x and drug y have similar potencies' or that 'drug x and drug y have similar durations of action' or maybe 'doses of drug x and drug y can be found where they have similar effects'. Of course, the required degree of similarity and the assays used for evaluation of the statistical hypothesis will have to be defined. Apart from the enormous differences in scope of the scientific and potential statistical hypotheses, the first may require several or all of the others to be true.
If you want to know if a hypothesis is a statistical hypothesis then if it concerns the value of a parameter within a statistical model or can be restated as being about a parameter value, then it is.
Now, the "alternative hypothesis". For the hypothesis testing framework there are two things that are commonly called 'alternative hypotheses'. The first is an arbitrary effect size that is used in the pre-data calculation of test power (usually for sample size determination). That alternative hypothesis is ONLY relevant before the data are in hand. Once you have the data the arbitrarily specified effect size loses its relevance because the observed effect size is known. When you perform the hypothesis test the effective alternative becomes nothing more than 'not the null'.
It is a bad mistake to assume that a rejection of the null hypothesis in a hypothesis test leads to the acceptance of the pre-data alternative hypothesis, and it is just about as bad to assume that it leads to the acceptance of the observed effect size as a true hypothesis.
Of course, the hypothesis test framework is not the only statistical approach, and I would argue, it is not even the most relevant to the majority of scientific endeavours. If you use a likelihood ratio test then you can compare the data support for two specified parameter values within the statistical model and that means that you can do the same within a Bayesian framework.
The principle of statistical hypothesis tests, by definition, treats the null hypothesis H0 and the alternative H1 asymmetrically. This always needs to be taken into account. A test is able to tell you whether there is evidence against the null hypothesis in the direction of the alternative.
It will never tell you that there is evidence against the alternative.
The choice of the H0 determines what the test can do; it determines what the test can indicate against.
I share @Michael Lew's reservations against a formal use of the term "proposed hypothesis", however let's assume for the following that you can translate your scientific research hypothesis into certain parameter values for a specified statistical model. Let's call this R.
If you choose R as H0, you can find evidence against it, but not in its favour. This may not be what you want - although it isn't out of question. You may well wonder whether certain data contradict your R, in which case you can use it as H0, however this has no potential, even in case of non-rejection, to convince other people that R is correct.
There is however a very reasonable scientific justification for using R as H0, which is that according to Popper in order to corroborate a scientific theory, you should try to falsify it, and the best corroboration comes from repeated attempts to falsify it (in a way in which it seems likely that the theory will be rejected if it is in fact false, which is what Mayo's "severity" concept is about). Apart from statistical error probabilities, this is what testing R as H0 actually allows to do, so there is a good reason for using R as H0.
If you choose R as H1, you can find evidence against the H0, which is not normally quite what you want, unless you interpret evidence against H0 as evidence in favour of your H1, which isn't necessarily granted (model assumptions may be violated for both H0 and H1, so they may both technically be wrong, and rejecting H0 doesn't "statistically prove" H1), although many would interpret a test in this way. It has value only to the extent that somebody who opposes your R argues that H0 might be true (as in "your hypothesised real effect does not exist, it's all just due to random variation"). In this case a test with R as H1 has at least the potential to indicate strongly against that H0. You can even go on and say it'll give you evidence that H0 is violated "in the direction of H1", but as said before there may be other explanations for this than that H1 is actually true. Also, "the direction of H1" is rather imprecise and doesn't amount to any specific parameter value or it's surroundings. It may depend on the application area how important that is. A homeopath may be happy enough to significantly show that homeopathy does something good rather than having its effect explained by random variation only, regardless of how much good it actually does, however precise numerical theories in physics/engineering, say, can hardly be backed up by just rejecting a random variation alternative.
The "equivalence testing" idea would amount to specifying a rather precise R (specific parameter value and small neighbourhood) as H1 and potentially rejecting a much bigger part of the parameter space on both sides of R. This would then be more informative, but has still the same issue with model assumptions, i.e., H0 and H1 may both be wrong. (Obviously model assumption diagnoses may help to some extent. Also even if neither H0 nor H1 is true, arguably some more distributions can be seen as "interpretatively equivalent" with them, e.g., two equal non-normal distributions in a two-sample setup where a normality-based test is applied, and actually may work well due to the Central Limit Theorem even for many non-normal distributions.)
So basically you need to choose what kind of statement you want to allow your test to back up. Choose R as H0 and the data can only reject it. Choose R as H1 and the data can reject the H0, and how valuable that is depends on the situation (particularly on how realistic the H0 looks as a competitor; i.e., how informative it actually is to reject it). The equivalence test setup is special by allowing you to use a rather precise R as H1 and reject a big H0, so the difference between this and rejecting a "random variation/no effect" H0 regards the precise or imprecise nature of the research hypothesis R to be tested.