Covariance matrix of data projected onto eigenvectors is diagonal.












1












$begingroup$


I am reading about PCA and found an exercise that says




Show that when a $N$-dim set of data points $X$ is projected onto the eigenvectors $V = [e_1 e_2...e_n]$ of its covariance matrix $C=XX^T$, the covariance matrix of the projected data $C_p = YY^T$ is diagonal and hence that, in the space of the eigenvector decomposition, the distribution of X is uncorrelated.




What I have so far is



$$Y = V^TX$$



Therefore



$$C_p = YY^T = V^TX(V^TX)^T = V^TXX^TV$$



but there I got stuck. Any advise on how to proceed, moreover, what does "The covariance matrix of the projected data is diagonal" mean?










share|cite|improve this question









$endgroup$

















    1












    $begingroup$


    I am reading about PCA and found an exercise that says




    Show that when a $N$-dim set of data points $X$ is projected onto the eigenvectors $V = [e_1 e_2...e_n]$ of its covariance matrix $C=XX^T$, the covariance matrix of the projected data $C_p = YY^T$ is diagonal and hence that, in the space of the eigenvector decomposition, the distribution of X is uncorrelated.




    What I have so far is



    $$Y = V^TX$$



    Therefore



    $$C_p = YY^T = V^TX(V^TX)^T = V^TXX^TV$$



    but there I got stuck. Any advise on how to proceed, moreover, what does "The covariance matrix of the projected data is diagonal" mean?










    share|cite|improve this question









    $endgroup$















      1












      1








      1


      1



      $begingroup$


      I am reading about PCA and found an exercise that says




      Show that when a $N$-dim set of data points $X$ is projected onto the eigenvectors $V = [e_1 e_2...e_n]$ of its covariance matrix $C=XX^T$, the covariance matrix of the projected data $C_p = YY^T$ is diagonal and hence that, in the space of the eigenvector decomposition, the distribution of X is uncorrelated.




      What I have so far is



      $$Y = V^TX$$



      Therefore



      $$C_p = YY^T = V^TX(V^TX)^T = V^TXX^TV$$



      but there I got stuck. Any advise on how to proceed, moreover, what does "The covariance matrix of the projected data is diagonal" mean?










      share|cite|improve this question









      $endgroup$




      I am reading about PCA and found an exercise that says




      Show that when a $N$-dim set of data points $X$ is projected onto the eigenvectors $V = [e_1 e_2...e_n]$ of its covariance matrix $C=XX^T$, the covariance matrix of the projected data $C_p = YY^T$ is diagonal and hence that, in the space of the eigenvector decomposition, the distribution of X is uncorrelated.




      What I have so far is



      $$Y = V^TX$$



      Therefore



      $$C_p = YY^T = V^TX(V^TX)^T = V^TXX^TV$$



      but there I got stuck. Any advise on how to proceed, moreover, what does "The covariance matrix of the projected data is diagonal" mean?







      linear-algebra






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Jan 18 '14 at 10:18









      BRabbit27BRabbit27

      352926




      352926






















          1 Answer
          1






          active

          oldest

          votes


















          0












          $begingroup$

          A diagonal matrix is one with zero everywhere and the diagonal entries can be zero or non zero.



          Assuming the eigenvectors are normalized in magnitude $|e_i|=1$
          $$
          begin{align}
          because text{ } & V = [e_1 e_2 dots e_n] text{ and } Ce_i =lambda_ie_i \\
          therefore text{ } &CV = V
          begin{bmatrix}
          lambda_1 & 0 & cdots & 0\
          0 & lambda_2 & cdots & 0\
          vdots & vdots & ddots & vdots\
          0 & 0 & cdots & lambda_n
          end{bmatrix}\\
          because text{ } & C_p = V^T (XX^T) V = V^TCV\\
          therefore text{ } & C_p=V^TVbegin{bmatrix}
          lambda_1 & 0 & cdots & 0\
          0 & lambda_2 & cdots & 0\
          vdots & vdots & ddots & vdots\
          0 & 0 & cdots & lambda_n
          end{bmatrix}\
          because text{ } & V^TV = mathbf{I}\
          &text{ as eigenvecotrs of a symmetric matrix are orthogonal}\\
          therefore text{ } &C_p= begin{bmatrix}
          lambda_1 & 0 & cdots & 0\
          0 & lambda_2 & cdots & 0\
          vdots & vdots & ddots & vdots\
          0 & 0 & cdots & lambda_n
          end{bmatrix}\
          &text{A diagonal matrix with variance in each eigenvector direction}
          end{align}
          $$






          share|cite|improve this answer









          $endgroup$













            Your Answer





            StackExchange.ifUsing("editor", function () {
            return StackExchange.using("mathjaxEditing", function () {
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            });
            });
            }, "mathjax-editing");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "69"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f642476%2fcovariance-matrix-of-data-projected-onto-eigenvectors-is-diagonal%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            1 Answer
            1






            active

            oldest

            votes








            1 Answer
            1






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            0












            $begingroup$

            A diagonal matrix is one with zero everywhere and the diagonal entries can be zero or non zero.



            Assuming the eigenvectors are normalized in magnitude $|e_i|=1$
            $$
            begin{align}
            because text{ } & V = [e_1 e_2 dots e_n] text{ and } Ce_i =lambda_ie_i \\
            therefore text{ } &CV = V
            begin{bmatrix}
            lambda_1 & 0 & cdots & 0\
            0 & lambda_2 & cdots & 0\
            vdots & vdots & ddots & vdots\
            0 & 0 & cdots & lambda_n
            end{bmatrix}\\
            because text{ } & C_p = V^T (XX^T) V = V^TCV\\
            therefore text{ } & C_p=V^TVbegin{bmatrix}
            lambda_1 & 0 & cdots & 0\
            0 & lambda_2 & cdots & 0\
            vdots & vdots & ddots & vdots\
            0 & 0 & cdots & lambda_n
            end{bmatrix}\
            because text{ } & V^TV = mathbf{I}\
            &text{ as eigenvecotrs of a symmetric matrix are orthogonal}\\
            therefore text{ } &C_p= begin{bmatrix}
            lambda_1 & 0 & cdots & 0\
            0 & lambda_2 & cdots & 0\
            vdots & vdots & ddots & vdots\
            0 & 0 & cdots & lambda_n
            end{bmatrix}\
            &text{A diagonal matrix with variance in each eigenvector direction}
            end{align}
            $$






            share|cite|improve this answer









            $endgroup$


















              0












              $begingroup$

              A diagonal matrix is one with zero everywhere and the diagonal entries can be zero or non zero.



              Assuming the eigenvectors are normalized in magnitude $|e_i|=1$
              $$
              begin{align}
              because text{ } & V = [e_1 e_2 dots e_n] text{ and } Ce_i =lambda_ie_i \\
              therefore text{ } &CV = V
              begin{bmatrix}
              lambda_1 & 0 & cdots & 0\
              0 & lambda_2 & cdots & 0\
              vdots & vdots & ddots & vdots\
              0 & 0 & cdots & lambda_n
              end{bmatrix}\\
              because text{ } & C_p = V^T (XX^T) V = V^TCV\\
              therefore text{ } & C_p=V^TVbegin{bmatrix}
              lambda_1 & 0 & cdots & 0\
              0 & lambda_2 & cdots & 0\
              vdots & vdots & ddots & vdots\
              0 & 0 & cdots & lambda_n
              end{bmatrix}\
              because text{ } & V^TV = mathbf{I}\
              &text{ as eigenvecotrs of a symmetric matrix are orthogonal}\\
              therefore text{ } &C_p= begin{bmatrix}
              lambda_1 & 0 & cdots & 0\
              0 & lambda_2 & cdots & 0\
              vdots & vdots & ddots & vdots\
              0 & 0 & cdots & lambda_n
              end{bmatrix}\
              &text{A diagonal matrix with variance in each eigenvector direction}
              end{align}
              $$






              share|cite|improve this answer









              $endgroup$
















                0












                0








                0





                $begingroup$

                A diagonal matrix is one with zero everywhere and the diagonal entries can be zero or non zero.



                Assuming the eigenvectors are normalized in magnitude $|e_i|=1$
                $$
                begin{align}
                because text{ } & V = [e_1 e_2 dots e_n] text{ and } Ce_i =lambda_ie_i \\
                therefore text{ } &CV = V
                begin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\\
                because text{ } & C_p = V^T (XX^T) V = V^TCV\\
                therefore text{ } & C_p=V^TVbegin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\
                because text{ } & V^TV = mathbf{I}\
                &text{ as eigenvecotrs of a symmetric matrix are orthogonal}\\
                therefore text{ } &C_p= begin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\
                &text{A diagonal matrix with variance in each eigenvector direction}
                end{align}
                $$






                share|cite|improve this answer









                $endgroup$



                A diagonal matrix is one with zero everywhere and the diagonal entries can be zero or non zero.



                Assuming the eigenvectors are normalized in magnitude $|e_i|=1$
                $$
                begin{align}
                because text{ } & V = [e_1 e_2 dots e_n] text{ and } Ce_i =lambda_ie_i \\
                therefore text{ } &CV = V
                begin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\\
                because text{ } & C_p = V^T (XX^T) V = V^TCV\\
                therefore text{ } & C_p=V^TVbegin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\
                because text{ } & V^TV = mathbf{I}\
                &text{ as eigenvecotrs of a symmetric matrix are orthogonal}\\
                therefore text{ } &C_p= begin{bmatrix}
                lambda_1 & 0 & cdots & 0\
                0 & lambda_2 & cdots & 0\
                vdots & vdots & ddots & vdots\
                0 & 0 & cdots & lambda_n
                end{bmatrix}\
                &text{A diagonal matrix with variance in each eigenvector direction}
                end{align}
                $$







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered Sep 28 '15 at 2:41









                Karim TarabishyKarim Tarabishy

                1206




                1206






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Mathematics Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f642476%2fcovariance-matrix-of-data-projected-onto-eigenvectors-is-diagonal%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    'app-layout' is not a known element: how to share Component with different Modules

                    android studio warns about leanback feature tag usage required on manifest while using Unity exported app?

                    SQL update select statement