Which test statistic is more effective in testing the variance of a normal variable?












0












$begingroup$


Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.




$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.




If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.




Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.




This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.



Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.



Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?



Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?










share|cite|improve this question











$endgroup$












  • $begingroup$
    Cross-posted here.
    $endgroup$
    – StubbornAtom
    Feb 3 at 7:07
















0












$begingroup$


Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.




$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.




If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.




Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.




This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.



Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.



Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?



Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?










share|cite|improve this question











$endgroup$












  • $begingroup$
    Cross-posted here.
    $endgroup$
    – StubbornAtom
    Feb 3 at 7:07














0












0








0


0



$begingroup$


Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.




$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.




If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.




Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.




This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.



Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.



Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?



Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?










share|cite|improve this question











$endgroup$




Most of the time, to do hypothesis tests is to follow the tedious processes given on statistics textbooks. But recently I find that there are, in general, more than several test statistics that can be used to test the same hypothesis. To give an example consider the following test.




$n$ samples of the random variable $X$ are given. Null Hypothesis $H_0:Xsim N(mu,sigma_0^2)$. Alternative Hypothesis $H_1:Xsim N(mu,sigma_1^2)$. $mu$ and $sigma_i$ are known. $sigma_0<sigma_1$.




If we follow the textbooks, then we should use the $chi$-square distribution, and the test statistic is the variance of the sample(times $n$ divided by $sigma_0$). But the following test is also appropriate.




Let $bar{X}$ be the sample mean. If $|bar{X}-mu|<c$ then accept $H_0$. Otherwise accept $H_1$. $c$ is chosen according to the level of significance of the test, $alpha$.




This is an appropriate test because the condition $|bar{X}-mu|<c$ is more likely to be satisfied under $H_0$ than under $H_1$. (Many other test statistics are also appropriate for the same reason.) The test, however, has one shortcoming: the probability of making a type II error, $beta$, is independent of $n$, which means it doesn't decrease as sample size increases.



Therefore I wonder which test statistic is the best in this situation. There is some simple criteria of how good a test statistic is. In this case, for example, we can say that for $n=100$ and $alpha=0.05$, the smaller the minimum value of $sigma_1/sigma_0$ required to make $betaleq 0.05$, the better the test statistic.



Also, how can we make sure that the test given on the textbook is the best test, and there are no better alternatives?



Since this question might be too broad, I would be glad if anyone could give a specific example(like the one above) to illustrate the general idea. Also, are there any books or articles that discuss this topic in detail?







statistics random-variables hypothesis-testing






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Feb 1 at 11:15







Holding Arthur

















asked Feb 1 at 3:31









Holding ArthurHolding Arthur

1,555417




1,555417












  • $begingroup$
    Cross-posted here.
    $endgroup$
    – StubbornAtom
    Feb 3 at 7:07


















  • $begingroup$
    Cross-posted here.
    $endgroup$
    – StubbornAtom
    Feb 3 at 7:07
















$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07




$begingroup$
Cross-posted here.
$endgroup$
– StubbornAtom
Feb 3 at 7:07










1 Answer
1






active

oldest

votes


















1












$begingroup$

Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance



So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:




  • Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct


  • Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct



So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test



With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$



The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses






share|cite|improve this answer









$endgroup$














    Your Answer





    StackExchange.ifUsing("editor", function () {
    return StackExchange.using("mathjaxEditing", function () {
    StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
    StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
    });
    });
    }, "mathjax-editing");

    StackExchange.ready(function() {
    var channelOptions = {
    tags: "".split(" "),
    id: "69"
    };
    initTagRenderer("".split(" "), "".split(" "), channelOptions);

    StackExchange.using("externalEditor", function() {
    // Have to fire editor after snippets, if snippets enabled
    if (StackExchange.settings.snippets.snippetsEnabled) {
    StackExchange.using("snippets", function() {
    createEditor();
    });
    }
    else {
    createEditor();
    }
    });

    function createEditor() {
    StackExchange.prepareEditor({
    heartbeatType: 'answer',
    autoActivateHeartbeat: false,
    convertImagesToLinks: true,
    noModals: true,
    showLowRepImageUploadWarning: true,
    reputationToPostImages: 10,
    bindNavPrevention: true,
    postfix: "",
    imageUploader: {
    brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
    contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
    allowUrls: true
    },
    noCode: true, onDemand: true,
    discardSelector: ".discard-answer"
    ,immediatelyShowMarkdownHelp:true
    });


    }
    });














    draft saved

    draft discarded


















    StackExchange.ready(
    function () {
    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3095784%2fwhich-test-statistic-is-more-effective-in-testing-the-variance-of-a-normal-varia%23new-answer', 'question_page');
    }
    );

    Post as a guest















    Required, but never shown

























    1 Answer
    1






    active

    oldest

    votes








    1 Answer
    1






    active

    oldest

    votes









    active

    oldest

    votes






    active

    oldest

    votes









    1












    $begingroup$

    Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance



    So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:




    • Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct


    • Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct



    So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test



    With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$



    The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses






    share|cite|improve this answer









    $endgroup$


















      1












      $begingroup$

      Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance



      So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:




      • Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct


      • Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct



      So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test



      With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$



      The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses






      share|cite|improve this answer









      $endgroup$
















        1












        1








        1





        $begingroup$

        Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance



        So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:




        • Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct


        • Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct



        So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test



        With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$



        The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses






        share|cite|improve this answer









        $endgroup$



        Typically, with a simple null and alternative hypothesis, the aim is to keep the probability of wrongly rejecting the null hypothesis (a Type I error) below a particular value (the significance level of the test) and then minimise the probability of wrongly failing to reject the null hypothesis (a Type II error), i.e. maximising the power of the test. Some tests are more powerful than others for the same level of significance



        So let's take your suggested test statistics with some arbitrary specific numbers: $H_0: X_i sim mathcal N(100,4)$ against $H_1: X_i sim mathcal N(100,9)$ with a sample size of $10$ aiming for a significance level of $0.05$, and compare:




        • Test A based on $frac1{sigma^2} sumlimits_i (X_i-mu)^2 sim chi^2_n$: reject $H_0$ if $frac1{4} sumlimits_{i=1}^{10} (X_i-100)^2 gt 18.307$. This has a probability of rejection of about $0.6155$ if $H_1$ is correct but $0.05$ if $H_0$ is correct


        • Test B based on $ frac{sqrt{n}}{sigma}left(bar{X}-muright) sim mathcal N(0,1)$: reject $H_0$ if $frac{sqrt{10}}{sqrt{4}} bigg|bar{X}-100bigg| gt 1.96$. This has a probability of rejection of about $0.1913$ if $H_1$ is correct but $0.05$ if $H_0$ is correct



        So Test A is more than three times as powerful as Test B in this particular case (seen another way, Test A is less than half as likely to make Type II errors), and in that sense Test A is a better test



        With larger $n$ the distinction is even greater: with $n=100$ the power of Test A using $chi^2_{100}$ rises to about $0.9999$ while, as you say, the power of Test B stays constant, at about $0.1913$



        The Neyman–Pearson_lemma can point to the uniformly most powerful test for simple hypotheses







        share|cite|improve this answer












        share|cite|improve this answer



        share|cite|improve this answer










        answered Feb 2 at 11:29









        HenryHenry

        101k482170




        101k482170






























            draft saved

            draft discarded




















































            Thanks for contributing an answer to Mathematics Stack Exchange!


            • Please be sure to answer the question. Provide details and share your research!

            But avoid



            • Asking for help, clarification, or responding to other answers.

            • Making statements based on opinion; back them up with references or personal experience.


            Use MathJax to format equations. MathJax reference.


            To learn more, see our tips on writing great answers.




            draft saved


            draft discarded














            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3095784%2fwhich-test-statistic-is-more-effective-in-testing-the-variance-of-a-normal-varia%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown





















































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown

































            Required, but never shown














            Required, but never shown












            Required, but never shown







            Required, but never shown







            Popular posts from this blog

            MongoDB - Not Authorized To Execute Command

            How to fix TextFormField cause rebuild widget in Flutter

            in spring boot 2.1 many test slices are not allowed anymore due to multiple @BootstrapWith