Determinant of matrix times elementary matrix












0














I have been reading Linear Algebra Done Wrong and came across this Lemma:



Lemma 3.6. For a square matrix A and an elementary matrix E (of the
same size)



     det(AE) = (det A)(det E)


Proof: The proof can be done just by direct checking: determinants of
special matrices are easy to compute; right multiplication by an elementary matrix is a column operation, and effect of column operations on the
determinant is well known.



This can look like a lucky coincidence, that the determinants of elementary matrices agree with the corresponding column operations, but it is not
a coincidence at all.



Namely, for a column operation the corresponding elementary matrix
can be obtained from the identity matrix I by this column operation. So, its
determinant is 1 (determinant of I) times the effect of the column operation.
And that is all! It may be hard to realize at first, but the above paragraph is a complete and rigorous proof of the lemma!



I am having a very hard time understanding this proof any help would be great, thanks!










share|cite|improve this question



























    0














    I have been reading Linear Algebra Done Wrong and came across this Lemma:



    Lemma 3.6. For a square matrix A and an elementary matrix E (of the
    same size)



         det(AE) = (det A)(det E)


    Proof: The proof can be done just by direct checking: determinants of
    special matrices are easy to compute; right multiplication by an elementary matrix is a column operation, and effect of column operations on the
    determinant is well known.



    This can look like a lucky coincidence, that the determinants of elementary matrices agree with the corresponding column operations, but it is not
    a coincidence at all.



    Namely, for a column operation the corresponding elementary matrix
    can be obtained from the identity matrix I by this column operation. So, its
    determinant is 1 (determinant of I) times the effect of the column operation.
    And that is all! It may be hard to realize at first, but the above paragraph is a complete and rigorous proof of the lemma!



    I am having a very hard time understanding this proof any help would be great, thanks!










    share|cite|improve this question

























      0












      0








      0







      I have been reading Linear Algebra Done Wrong and came across this Lemma:



      Lemma 3.6. For a square matrix A and an elementary matrix E (of the
      same size)



           det(AE) = (det A)(det E)


      Proof: The proof can be done just by direct checking: determinants of
      special matrices are easy to compute; right multiplication by an elementary matrix is a column operation, and effect of column operations on the
      determinant is well known.



      This can look like a lucky coincidence, that the determinants of elementary matrices agree with the corresponding column operations, but it is not
      a coincidence at all.



      Namely, for a column operation the corresponding elementary matrix
      can be obtained from the identity matrix I by this column operation. So, its
      determinant is 1 (determinant of I) times the effect of the column operation.
      And that is all! It may be hard to realize at first, but the above paragraph is a complete and rigorous proof of the lemma!



      I am having a very hard time understanding this proof any help would be great, thanks!










      share|cite|improve this question













      I have been reading Linear Algebra Done Wrong and came across this Lemma:



      Lemma 3.6. For a square matrix A and an elementary matrix E (of the
      same size)



           det(AE) = (det A)(det E)


      Proof: The proof can be done just by direct checking: determinants of
      special matrices are easy to compute; right multiplication by an elementary matrix is a column operation, and effect of column operations on the
      determinant is well known.



      This can look like a lucky coincidence, that the determinants of elementary matrices agree with the corresponding column operations, but it is not
      a coincidence at all.



      Namely, for a column operation the corresponding elementary matrix
      can be obtained from the identity matrix I by this column operation. So, its
      determinant is 1 (determinant of I) times the effect of the column operation.
      And that is all! It may be hard to realize at first, but the above paragraph is a complete and rigorous proof of the lemma!



      I am having a very hard time understanding this proof any help would be great, thanks!







      linear-algebra determinant






      share|cite|improve this question













      share|cite|improve this question











      share|cite|improve this question




      share|cite|improve this question










      asked Jan 1 at 2:24









      jake walshjake walsh

      808




      808






















          2 Answers
          2






          active

          oldest

          votes


















          4














          OK, so this is a weird proof, but there's really only two parts of this proof that actually matter here. First:




          [...] right multiplication by an elementary matrix is a column operation, and effect of column operations on the determinant is well known.




          What this means is that when you multiply $A$ by $E$ on the right, you are doing a column operation on $A$, whether that be switching two columns, doing a column addition, or multiplying a column by a scalar. All of these operations will multiply the determinant by some known number, $k$ (i.e. switching two columns multiplies determinant by $k=-1$, column addition multiplies determinant by $k=1$, and multiplying a column by the scalar $c$ multiplies determinant by $k=c$). Thus, for any matrix $A$, the determinant of $AE$ will be $k$ times the original determinant of $A$. We can write this in an equation as follows:



          $$det AE=k(det A)$$



          Second:




          [...] for a column operation the corresponding elementary matrix can be obtained from the identity matrix I by this column operation. So, its determinant is 1 (determinant of I) times the effect of the column operation.




          Now, this is really confusing at first, but it can be understood in terms of our $det AE=k(det A)$ above. See, this equation works for any matrix $A$, which means we could also substitute the identity matrix $I$ for $A$ into this equation. Therefore, we get:



          $$det IE=k(det I)rightarrow det E=k$$



          (Note that this uses the fact that $IE=E$ and $det I=1$.)



          Therefore, we now know the value of $k$ is $det E$. Thus, we can substitute that back into our original equation $det AE=k(det A)$ to get:



          $$det AE=(det E)(det A)$$



          Since scalar multiplication is commutative, we can switch the right side around to get the final lemma:



          $$det AE=(det A)(det E)$$






          share|cite|improve this answer





















          • Brilliant thanks very much!
            – jake walsh
            Jan 1 at 2:41



















          0














          The text defines the determinant in terms of column operation and there is an elementary matrix for each column operation. They spend a lot of time building the determinant based on column operations and the payoff is that this lemma is essentially true by construction.






          share|cite|improve this answer





















            Your Answer





            StackExchange.ifUsing("editor", function () {
            return StackExchange.using("mathjaxEditing", function () {
            StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
            StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
            });
            });
            }, "mathjax-editing");

            StackExchange.ready(function() {
            var channelOptions = {
            tags: "".split(" "),
            id: "69"
            };
            initTagRenderer("".split(" "), "".split(" "), channelOptions);

            StackExchange.using("externalEditor", function() {
            // Have to fire editor after snippets, if snippets enabled
            if (StackExchange.settings.snippets.snippetsEnabled) {
            StackExchange.using("snippets", function() {
            createEditor();
            });
            }
            else {
            createEditor();
            }
            });

            function createEditor() {
            StackExchange.prepareEditor({
            heartbeatType: 'answer',
            autoActivateHeartbeat: false,
            convertImagesToLinks: true,
            noModals: true,
            showLowRepImageUploadWarning: true,
            reputationToPostImages: 10,
            bindNavPrevention: true,
            postfix: "",
            imageUploader: {
            brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
            contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
            allowUrls: true
            },
            noCode: true, onDemand: true,
            discardSelector: ".discard-answer"
            ,immediatelyShowMarkdownHelp:true
            });


            }
            });














            draft saved

            draft discarded


















            StackExchange.ready(
            function () {
            StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3058178%2fdeterminant-of-matrix-times-elementary-matrix%23new-answer', 'question_page');
            }
            );

            Post as a guest















            Required, but never shown

























            2 Answers
            2






            active

            oldest

            votes








            2 Answers
            2






            active

            oldest

            votes









            active

            oldest

            votes






            active

            oldest

            votes









            4














            OK, so this is a weird proof, but there's really only two parts of this proof that actually matter here. First:




            [...] right multiplication by an elementary matrix is a column operation, and effect of column operations on the determinant is well known.




            What this means is that when you multiply $A$ by $E$ on the right, you are doing a column operation on $A$, whether that be switching two columns, doing a column addition, or multiplying a column by a scalar. All of these operations will multiply the determinant by some known number, $k$ (i.e. switching two columns multiplies determinant by $k=-1$, column addition multiplies determinant by $k=1$, and multiplying a column by the scalar $c$ multiplies determinant by $k=c$). Thus, for any matrix $A$, the determinant of $AE$ will be $k$ times the original determinant of $A$. We can write this in an equation as follows:



            $$det AE=k(det A)$$



            Second:




            [...] for a column operation the corresponding elementary matrix can be obtained from the identity matrix I by this column operation. So, its determinant is 1 (determinant of I) times the effect of the column operation.




            Now, this is really confusing at first, but it can be understood in terms of our $det AE=k(det A)$ above. See, this equation works for any matrix $A$, which means we could also substitute the identity matrix $I$ for $A$ into this equation. Therefore, we get:



            $$det IE=k(det I)rightarrow det E=k$$



            (Note that this uses the fact that $IE=E$ and $det I=1$.)



            Therefore, we now know the value of $k$ is $det E$. Thus, we can substitute that back into our original equation $det AE=k(det A)$ to get:



            $$det AE=(det E)(det A)$$



            Since scalar multiplication is commutative, we can switch the right side around to get the final lemma:



            $$det AE=(det A)(det E)$$






            share|cite|improve this answer





















            • Brilliant thanks very much!
              – jake walsh
              Jan 1 at 2:41
















            4














            OK, so this is a weird proof, but there's really only two parts of this proof that actually matter here. First:




            [...] right multiplication by an elementary matrix is a column operation, and effect of column operations on the determinant is well known.




            What this means is that when you multiply $A$ by $E$ on the right, you are doing a column operation on $A$, whether that be switching two columns, doing a column addition, or multiplying a column by a scalar. All of these operations will multiply the determinant by some known number, $k$ (i.e. switching two columns multiplies determinant by $k=-1$, column addition multiplies determinant by $k=1$, and multiplying a column by the scalar $c$ multiplies determinant by $k=c$). Thus, for any matrix $A$, the determinant of $AE$ will be $k$ times the original determinant of $A$. We can write this in an equation as follows:



            $$det AE=k(det A)$$



            Second:




            [...] for a column operation the corresponding elementary matrix can be obtained from the identity matrix I by this column operation. So, its determinant is 1 (determinant of I) times the effect of the column operation.




            Now, this is really confusing at first, but it can be understood in terms of our $det AE=k(det A)$ above. See, this equation works for any matrix $A$, which means we could also substitute the identity matrix $I$ for $A$ into this equation. Therefore, we get:



            $$det IE=k(det I)rightarrow det E=k$$



            (Note that this uses the fact that $IE=E$ and $det I=1$.)



            Therefore, we now know the value of $k$ is $det E$. Thus, we can substitute that back into our original equation $det AE=k(det A)$ to get:



            $$det AE=(det E)(det A)$$



            Since scalar multiplication is commutative, we can switch the right side around to get the final lemma:



            $$det AE=(det A)(det E)$$






            share|cite|improve this answer





















            • Brilliant thanks very much!
              – jake walsh
              Jan 1 at 2:41














            4












            4








            4






            OK, so this is a weird proof, but there's really only two parts of this proof that actually matter here. First:




            [...] right multiplication by an elementary matrix is a column operation, and effect of column operations on the determinant is well known.




            What this means is that when you multiply $A$ by $E$ on the right, you are doing a column operation on $A$, whether that be switching two columns, doing a column addition, or multiplying a column by a scalar. All of these operations will multiply the determinant by some known number, $k$ (i.e. switching two columns multiplies determinant by $k=-1$, column addition multiplies determinant by $k=1$, and multiplying a column by the scalar $c$ multiplies determinant by $k=c$). Thus, for any matrix $A$, the determinant of $AE$ will be $k$ times the original determinant of $A$. We can write this in an equation as follows:



            $$det AE=k(det A)$$



            Second:




            [...] for a column operation the corresponding elementary matrix can be obtained from the identity matrix I by this column operation. So, its determinant is 1 (determinant of I) times the effect of the column operation.




            Now, this is really confusing at first, but it can be understood in terms of our $det AE=k(det A)$ above. See, this equation works for any matrix $A$, which means we could also substitute the identity matrix $I$ for $A$ into this equation. Therefore, we get:



            $$det IE=k(det I)rightarrow det E=k$$



            (Note that this uses the fact that $IE=E$ and $det I=1$.)



            Therefore, we now know the value of $k$ is $det E$. Thus, we can substitute that back into our original equation $det AE=k(det A)$ to get:



            $$det AE=(det E)(det A)$$



            Since scalar multiplication is commutative, we can switch the right side around to get the final lemma:



            $$det AE=(det A)(det E)$$






            share|cite|improve this answer












            OK, so this is a weird proof, but there's really only two parts of this proof that actually matter here. First:




            [...] right multiplication by an elementary matrix is a column operation, and effect of column operations on the determinant is well known.




            What this means is that when you multiply $A$ by $E$ on the right, you are doing a column operation on $A$, whether that be switching two columns, doing a column addition, or multiplying a column by a scalar. All of these operations will multiply the determinant by some known number, $k$ (i.e. switching two columns multiplies determinant by $k=-1$, column addition multiplies determinant by $k=1$, and multiplying a column by the scalar $c$ multiplies determinant by $k=c$). Thus, for any matrix $A$, the determinant of $AE$ will be $k$ times the original determinant of $A$. We can write this in an equation as follows:



            $$det AE=k(det A)$$



            Second:




            [...] for a column operation the corresponding elementary matrix can be obtained from the identity matrix I by this column operation. So, its determinant is 1 (determinant of I) times the effect of the column operation.




            Now, this is really confusing at first, but it can be understood in terms of our $det AE=k(det A)$ above. See, this equation works for any matrix $A$, which means we could also substitute the identity matrix $I$ for $A$ into this equation. Therefore, we get:



            $$det IE=k(det I)rightarrow det E=k$$



            (Note that this uses the fact that $IE=E$ and $det I=1$.)



            Therefore, we now know the value of $k$ is $det E$. Thus, we can substitute that back into our original equation $det AE=k(det A)$ to get:



            $$det AE=(det E)(det A)$$



            Since scalar multiplication is commutative, we can switch the right side around to get the final lemma:



            $$det AE=(det A)(det E)$$







            share|cite|improve this answer












            share|cite|improve this answer



            share|cite|improve this answer










            answered Jan 1 at 2:38









            Noble MushtakNoble Mushtak

            15.2k1735




            15.2k1735












            • Brilliant thanks very much!
              – jake walsh
              Jan 1 at 2:41


















            • Brilliant thanks very much!
              – jake walsh
              Jan 1 at 2:41
















            Brilliant thanks very much!
            – jake walsh
            Jan 1 at 2:41




            Brilliant thanks very much!
            – jake walsh
            Jan 1 at 2:41











            0














            The text defines the determinant in terms of column operation and there is an elementary matrix for each column operation. They spend a lot of time building the determinant based on column operations and the payoff is that this lemma is essentially true by construction.






            share|cite|improve this answer


























              0














              The text defines the determinant in terms of column operation and there is an elementary matrix for each column operation. They spend a lot of time building the determinant based on column operations and the payoff is that this lemma is essentially true by construction.






              share|cite|improve this answer
























                0












                0








                0






                The text defines the determinant in terms of column operation and there is an elementary matrix for each column operation. They spend a lot of time building the determinant based on column operations and the payoff is that this lemma is essentially true by construction.






                share|cite|improve this answer












                The text defines the determinant in terms of column operation and there is an elementary matrix for each column operation. They spend a lot of time building the determinant based on column operations and the payoff is that this lemma is essentially true by construction.







                share|cite|improve this answer












                share|cite|improve this answer



                share|cite|improve this answer










                answered Jan 1 at 2:42









                CyclotomicFieldCyclotomicField

                2,1921313




                2,1921313






























                    draft saved

                    draft discarded




















































                    Thanks for contributing an answer to Mathematics Stack Exchange!


                    • Please be sure to answer the question. Provide details and share your research!

                    But avoid



                    • Asking for help, clarification, or responding to other answers.

                    • Making statements based on opinion; back them up with references or personal experience.


                    Use MathJax to format equations. MathJax reference.


                    To learn more, see our tips on writing great answers.




                    draft saved


                    draft discarded














                    StackExchange.ready(
                    function () {
                    StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3058178%2fdeterminant-of-matrix-times-elementary-matrix%23new-answer', 'question_page');
                    }
                    );

                    Post as a guest















                    Required, but never shown





















































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown

































                    Required, but never shown














                    Required, but never shown












                    Required, but never shown







                    Required, but never shown







                    Popular posts from this blog

                    MongoDB - Not Authorized To Execute Command

                    How to fix TextFormField cause rebuild widget in Flutter

                    in spring boot 2.1 many test slices are not allowed anymore due to multiple @BootstrapWith