Which test statistic is more effective in testing the variance of a normal variable?
$begingroup$
Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.
$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.
If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.
Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.
This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.
Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.
Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?
Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?
statistics random-variables hypothesis-testing
$endgroup$
add a comment |
$begingroup$
Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.
$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.
If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.
Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.
This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.
Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.
Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?
Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?
statistics random-variables hypothesis-testing
$endgroup$
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07
add a comment |
$begingroup$
Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.
$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.
If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.
Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.
This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.
Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.
Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?
Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?
statistics random-variables hypothesis-testing
$endgroup$
Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.
$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.
If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.
Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.
This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.
Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.
Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?
Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?
statistics random-variables hypothesis-testing
statistics random-variables hypothesis-testing
edited Feb 1 at 11:15
Holding Arthur
asked Feb 1 at 3:31


Holding ArthurHolding Arthur
1,555417
1,555417
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07
add a comment |
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07
add a comment |
1 Answer
1
active
oldest
votes
$begingroup$
Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance
So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:
Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test
With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$
The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3095784%2fwhich-test-statistic-is-more-effective-in-testing-the-variance-of-a-normal-varia%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance
So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:
Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test
With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$
The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses
$endgroup$
add a comment |
$begingroup$
Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance
So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:
Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test
With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$
The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses
$endgroup$
add a comment |
$begingroup$
Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance
So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:
Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test
With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$
The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses
$endgroup$
Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance
So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:
Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct
So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test
With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$
The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses
answered Feb 2 at 11:29
HenryHenry
101k482170
101k482170
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3095784%2fwhich-test-statistic-is-more-effective-in-testing-the-variance-of-a-normal-varia%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07