Simple proof of the following: A matrix $A$ is onto if and only if its rows are linearly independent












0














Saw this claim in this video https://www.youtube.com/watch?v=cBSf2pGYAcA&list=PL06960BA52D0DB32B&index=4 at 23m35s and struggled in vain to prove it.



Searching for it on Google proved fruitless.










share|cite|improve this question
























  • What do you mean by "matrix $A$ is onto"?
    – Somos
    Nov 20 '18 at 1:35












  • $$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
    – James Shapiro
    Nov 20 '18 at 1:49


















0














Saw this claim in this video https://www.youtube.com/watch?v=cBSf2pGYAcA&list=PL06960BA52D0DB32B&index=4 at 23m35s and struggled in vain to prove it.



Searching for it on Google proved fruitless.










share|cite|improve this question
























  • What do you mean by "matrix $A$ is onto"?
    – Somos
    Nov 20 '18 at 1:35












  • $$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
    – James Shapiro
    Nov 20 '18 at 1:49
















0












0








0


1





Saw this claim in this video https://www.youtube.com/watch?v=cBSf2pGYAcA&list=PL06960BA52D0DB32B&index=4 at 23m35s and struggled in vain to prove it.



Searching for it on Google proved fruitless.










share|cite|improve this question















Saw this claim in this video https://www.youtube.com/watch?v=cBSf2pGYAcA&list=PL06960BA52D0DB32B&index=4 at 23m35s and struggled in vain to prove it.



Searching for it on Google proved fruitless.







linear-algebra






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Nov 20 '18 at 1:24









Tianlalu

3,09621038




3,09621038










asked Nov 20 '18 at 1:14









James Shapiro

1113




1113












  • What do you mean by "matrix $A$ is onto"?
    – Somos
    Nov 20 '18 at 1:35












  • $$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
    – James Shapiro
    Nov 20 '18 at 1:49




















  • What do you mean by "matrix $A$ is onto"?
    – Somos
    Nov 20 '18 at 1:35












  • $$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
    – James Shapiro
    Nov 20 '18 at 1:49


















What do you mean by "matrix $A$ is onto"?
– Somos
Nov 20 '18 at 1:35






What do you mean by "matrix $A$ is onto"?
– Somos
Nov 20 '18 at 1:35














$$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
– James Shapiro
Nov 20 '18 at 1:49






$$forall y in mathbb{R}^m, exists x textrm{such that} y = Ax$$
– James Shapiro
Nov 20 '18 at 1:49












2 Answers
2






active

oldest

votes


















2














Here's a simple proof. Let $V$ and $W$ be vector spaces. Let $T:Vto W$ be a linear map. Let $A$ be any matrix representing this map. Then $newcommandrk{operatorname{rank}}rk A$ is the dimension of the image of $T$, and the number of rows of $A$ is the dimension of $W$. Thus for $T$ to be surjective, we must have that the rank of $A$ is equal to the number of rows of $A$. Since row rank equals column rank, the rank of $A$ is the number of linearly independent rows of $A$. Hence $T$ is surjective if and only if the rows of any matrix representing it are linearly independent.



Edit I realized there a perhaps slightly more direct proof, which sort of mixes the proof above with the proof that row rank equals column rank.



Review of some relevant linear algebra



First recall some notions. If $V$ is a vector space over a field $K$, then the dual vector space to $V$ is defined to be $V^*:=newcommandHom{operatorname{Hom}}Hom_K(V,K)$. If $T:Vto W$ is a linear map, then its dual is $T^*:W^*to V^*$ defined by $T^*(lambda) = lambda circ T$. Also if ${v_i}$ and ${w_j}$ are bases for $V$ and $W$ with dual bases ${v_i^*}$ and ${w_j^*}$, then it is easy to show that the matrix for $T$ with respect to the $v_i$ and $w_j$s is $A_{ij}=w_j^*Tv_i$, and the matrix for $T^*$ with respect to $w_j^*$ and $v_i^*$ is $B_{ji}=(T^*w_j^*)v_i=(w_j^*circ T)v_i=w_j^*Tv_i=A_{ij}$. Thus the matrix for $T^*$ is the transpose of the matrix for $T$.



Now onto the proof



Proof



Let $T:Vto W$. Let $A$ be any matrix representing this map. Then $T$ is surjective if and only if there are no nonzero linear functionals on $W$ that vanish on the image of $T$. However if $lambda$ vanishes on the image of $T$, then we have $T^*lambda=0$, and recalling that the matrix of $T^*$ is $A^t$, we see that the coordinates of $lambda$ give an explicit linear dependence among the columns of $A^t$ (which are the rows of $A$). Conversely any linear dependence among the rows of $A$ gives such a linear functional $lambda$, proving that $T$ is not surjective.






share|cite|improve this answer























  • I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
    – James Shapiro
    Nov 20 '18 at 1:35










  • @JamesShapiro That's also a reasonable direction of proof.
    – jgon
    Nov 20 '18 at 1:38



















0














For a matrix A to be onto, there has to be a pivot in every row. To test the linear independence of the rows, you can look at A$^T$ and row reduce. If every column of A$^T$ has a pivot, the columns are linearly independent. Note that the columns of A$^T$ are the rows of A. I leave it to you to show that the row reduction doesn't affect the linear independence.






share|cite|improve this answer





















  • Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
    – James Shapiro
    Nov 20 '18 at 1:24













Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3005793%2fsimple-proof-of-the-following-a-matrix-a-is-onto-if-and-only-if-its-rows-are%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























2 Answers
2






active

oldest

votes








2 Answers
2






active

oldest

votes









active

oldest

votes






active

oldest

votes









2














Here's a simple proof. Let $V$ and $W$ be vector spaces. Let $T:Vto W$ be a linear map. Let $A$ be any matrix representing this map. Then $newcommandrk{operatorname{rank}}rk A$ is the dimension of the image of $T$, and the number of rows of $A$ is the dimension of $W$. Thus for $T$ to be surjective, we must have that the rank of $A$ is equal to the number of rows of $A$. Since row rank equals column rank, the rank of $A$ is the number of linearly independent rows of $A$. Hence $T$ is surjective if and only if the rows of any matrix representing it are linearly independent.



Edit I realized there a perhaps slightly more direct proof, which sort of mixes the proof above with the proof that row rank equals column rank.



Review of some relevant linear algebra



First recall some notions. If $V$ is a vector space over a field $K$, then the dual vector space to $V$ is defined to be $V^*:=newcommandHom{operatorname{Hom}}Hom_K(V,K)$. If $T:Vto W$ is a linear map, then its dual is $T^*:W^*to V^*$ defined by $T^*(lambda) = lambda circ T$. Also if ${v_i}$ and ${w_j}$ are bases for $V$ and $W$ with dual bases ${v_i^*}$ and ${w_j^*}$, then it is easy to show that the matrix for $T$ with respect to the $v_i$ and $w_j$s is $A_{ij}=w_j^*Tv_i$, and the matrix for $T^*$ with respect to $w_j^*$ and $v_i^*$ is $B_{ji}=(T^*w_j^*)v_i=(w_j^*circ T)v_i=w_j^*Tv_i=A_{ij}$. Thus the matrix for $T^*$ is the transpose of the matrix for $T$.



Now onto the proof



Proof



Let $T:Vto W$. Let $A$ be any matrix representing this map. Then $T$ is surjective if and only if there are no nonzero linear functionals on $W$ that vanish on the image of $T$. However if $lambda$ vanishes on the image of $T$, then we have $T^*lambda=0$, and recalling that the matrix of $T^*$ is $A^t$, we see that the coordinates of $lambda$ give an explicit linear dependence among the columns of $A^t$ (which are the rows of $A$). Conversely any linear dependence among the rows of $A$ gives such a linear functional $lambda$, proving that $T$ is not surjective.






share|cite|improve this answer























  • I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
    – James Shapiro
    Nov 20 '18 at 1:35










  • @JamesShapiro That's also a reasonable direction of proof.
    – jgon
    Nov 20 '18 at 1:38
















2














Here's a simple proof. Let $V$ and $W$ be vector spaces. Let $T:Vto W$ be a linear map. Let $A$ be any matrix representing this map. Then $newcommandrk{operatorname{rank}}rk A$ is the dimension of the image of $T$, and the number of rows of $A$ is the dimension of $W$. Thus for $T$ to be surjective, we must have that the rank of $A$ is equal to the number of rows of $A$. Since row rank equals column rank, the rank of $A$ is the number of linearly independent rows of $A$. Hence $T$ is surjective if and only if the rows of any matrix representing it are linearly independent.



Edit I realized there a perhaps slightly more direct proof, which sort of mixes the proof above with the proof that row rank equals column rank.



Review of some relevant linear algebra



First recall some notions. If $V$ is a vector space over a field $K$, then the dual vector space to $V$ is defined to be $V^*:=newcommandHom{operatorname{Hom}}Hom_K(V,K)$. If $T:Vto W$ is a linear map, then its dual is $T^*:W^*to V^*$ defined by $T^*(lambda) = lambda circ T$. Also if ${v_i}$ and ${w_j}$ are bases for $V$ and $W$ with dual bases ${v_i^*}$ and ${w_j^*}$, then it is easy to show that the matrix for $T$ with respect to the $v_i$ and $w_j$s is $A_{ij}=w_j^*Tv_i$, and the matrix for $T^*$ with respect to $w_j^*$ and $v_i^*$ is $B_{ji}=(T^*w_j^*)v_i=(w_j^*circ T)v_i=w_j^*Tv_i=A_{ij}$. Thus the matrix for $T^*$ is the transpose of the matrix for $T$.



Now onto the proof



Proof



Let $T:Vto W$. Let $A$ be any matrix representing this map. Then $T$ is surjective if and only if there are no nonzero linear functionals on $W$ that vanish on the image of $T$. However if $lambda$ vanishes on the image of $T$, then we have $T^*lambda=0$, and recalling that the matrix of $T^*$ is $A^t$, we see that the coordinates of $lambda$ give an explicit linear dependence among the columns of $A^t$ (which are the rows of $A$). Conversely any linear dependence among the rows of $A$ gives such a linear functional $lambda$, proving that $T$ is not surjective.






share|cite|improve this answer























  • I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
    – James Shapiro
    Nov 20 '18 at 1:35










  • @JamesShapiro That's also a reasonable direction of proof.
    – jgon
    Nov 20 '18 at 1:38














2












2








2






Here's a simple proof. Let $V$ and $W$ be vector spaces. Let $T:Vto W$ be a linear map. Let $A$ be any matrix representing this map. Then $newcommandrk{operatorname{rank}}rk A$ is the dimension of the image of $T$, and the number of rows of $A$ is the dimension of $W$. Thus for $T$ to be surjective, we must have that the rank of $A$ is equal to the number of rows of $A$. Since row rank equals column rank, the rank of $A$ is the number of linearly independent rows of $A$. Hence $T$ is surjective if and only if the rows of any matrix representing it are linearly independent.



Edit I realized there a perhaps slightly more direct proof, which sort of mixes the proof above with the proof that row rank equals column rank.



Review of some relevant linear algebra



First recall some notions. If $V$ is a vector space over a field $K$, then the dual vector space to $V$ is defined to be $V^*:=newcommandHom{operatorname{Hom}}Hom_K(V,K)$. If $T:Vto W$ is a linear map, then its dual is $T^*:W^*to V^*$ defined by $T^*(lambda) = lambda circ T$. Also if ${v_i}$ and ${w_j}$ are bases for $V$ and $W$ with dual bases ${v_i^*}$ and ${w_j^*}$, then it is easy to show that the matrix for $T$ with respect to the $v_i$ and $w_j$s is $A_{ij}=w_j^*Tv_i$, and the matrix for $T^*$ with respect to $w_j^*$ and $v_i^*$ is $B_{ji}=(T^*w_j^*)v_i=(w_j^*circ T)v_i=w_j^*Tv_i=A_{ij}$. Thus the matrix for $T^*$ is the transpose of the matrix for $T$.



Now onto the proof



Proof



Let $T:Vto W$. Let $A$ be any matrix representing this map. Then $T$ is surjective if and only if there are no nonzero linear functionals on $W$ that vanish on the image of $T$. However if $lambda$ vanishes on the image of $T$, then we have $T^*lambda=0$, and recalling that the matrix of $T^*$ is $A^t$, we see that the coordinates of $lambda$ give an explicit linear dependence among the columns of $A^t$ (which are the rows of $A$). Conversely any linear dependence among the rows of $A$ gives such a linear functional $lambda$, proving that $T$ is not surjective.






share|cite|improve this answer














Here's a simple proof. Let $V$ and $W$ be vector spaces. Let $T:Vto W$ be a linear map. Let $A$ be any matrix representing this map. Then $newcommandrk{operatorname{rank}}rk A$ is the dimension of the image of $T$, and the number of rows of $A$ is the dimension of $W$. Thus for $T$ to be surjective, we must have that the rank of $A$ is equal to the number of rows of $A$. Since row rank equals column rank, the rank of $A$ is the number of linearly independent rows of $A$. Hence $T$ is surjective if and only if the rows of any matrix representing it are linearly independent.



Edit I realized there a perhaps slightly more direct proof, which sort of mixes the proof above with the proof that row rank equals column rank.



Review of some relevant linear algebra



First recall some notions. If $V$ is a vector space over a field $K$, then the dual vector space to $V$ is defined to be $V^*:=newcommandHom{operatorname{Hom}}Hom_K(V,K)$. If $T:Vto W$ is a linear map, then its dual is $T^*:W^*to V^*$ defined by $T^*(lambda) = lambda circ T$. Also if ${v_i}$ and ${w_j}$ are bases for $V$ and $W$ with dual bases ${v_i^*}$ and ${w_j^*}$, then it is easy to show that the matrix for $T$ with respect to the $v_i$ and $w_j$s is $A_{ij}=w_j^*Tv_i$, and the matrix for $T^*$ with respect to $w_j^*$ and $v_i^*$ is $B_{ji}=(T^*w_j^*)v_i=(w_j^*circ T)v_i=w_j^*Tv_i=A_{ij}$. Thus the matrix for $T^*$ is the transpose of the matrix for $T$.



Now onto the proof



Proof



Let $T:Vto W$. Let $A$ be any matrix representing this map. Then $T$ is surjective if and only if there are no nonzero linear functionals on $W$ that vanish on the image of $T$. However if $lambda$ vanishes on the image of $T$, then we have $T^*lambda=0$, and recalling that the matrix of $T^*$ is $A^t$, we see that the coordinates of $lambda$ give an explicit linear dependence among the columns of $A^t$ (which are the rows of $A$). Conversely any linear dependence among the rows of $A$ gives such a linear functional $lambda$, proving that $T$ is not surjective.







share|cite|improve this answer














share|cite|improve this answer



share|cite|improve this answer








edited Nov 20 '18 at 23:38

























answered Nov 20 '18 at 1:29









jgon

12.8k21940




12.8k21940












  • I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
    – James Shapiro
    Nov 20 '18 at 1:35










  • @JamesShapiro That's also a reasonable direction of proof.
    – jgon
    Nov 20 '18 at 1:38


















  • I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
    – James Shapiro
    Nov 20 '18 at 1:35










  • @JamesShapiro That's also a reasonable direction of proof.
    – jgon
    Nov 20 '18 at 1:38
















I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
– James Shapiro
Nov 20 '18 at 1:35




I think that works. It might also be worth noting that for an m x n matrix to be surjective, we must have that m <= n (i.e. it must be a fat matrix). Hence, the rank of the matrix is at most min(m, n) = m. And if the rows are linearly dependent, then we have that the rank is strictly less than m.
– James Shapiro
Nov 20 '18 at 1:35












@JamesShapiro That's also a reasonable direction of proof.
– jgon
Nov 20 '18 at 1:38




@JamesShapiro That's also a reasonable direction of proof.
– jgon
Nov 20 '18 at 1:38











0














For a matrix A to be onto, there has to be a pivot in every row. To test the linear independence of the rows, you can look at A$^T$ and row reduce. If every column of A$^T$ has a pivot, the columns are linearly independent. Note that the columns of A$^T$ are the rows of A. I leave it to you to show that the row reduction doesn't affect the linear independence.






share|cite|improve this answer





















  • Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
    – James Shapiro
    Nov 20 '18 at 1:24


















0














For a matrix A to be onto, there has to be a pivot in every row. To test the linear independence of the rows, you can look at A$^T$ and row reduce. If every column of A$^T$ has a pivot, the columns are linearly independent. Note that the columns of A$^T$ are the rows of A. I leave it to you to show that the row reduction doesn't affect the linear independence.






share|cite|improve this answer





















  • Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
    – James Shapiro
    Nov 20 '18 at 1:24
















0












0








0






For a matrix A to be onto, there has to be a pivot in every row. To test the linear independence of the rows, you can look at A$^T$ and row reduce. If every column of A$^T$ has a pivot, the columns are linearly independent. Note that the columns of A$^T$ are the rows of A. I leave it to you to show that the row reduction doesn't affect the linear independence.






share|cite|improve this answer












For a matrix A to be onto, there has to be a pivot in every row. To test the linear independence of the rows, you can look at A$^T$ and row reduce. If every column of A$^T$ has a pivot, the columns are linearly independent. Note that the columns of A$^T$ are the rows of A. I leave it to you to show that the row reduction doesn't affect the linear independence.







share|cite|improve this answer












share|cite|improve this answer



share|cite|improve this answer










answered Nov 20 '18 at 1:23









Joel Pereira

65119




65119












  • Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
    – James Shapiro
    Nov 20 '18 at 1:24




















  • Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
    – James Shapiro
    Nov 20 '18 at 1:24


















Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
– James Shapiro
Nov 20 '18 at 1:24






Okay, I follow that argument and was aware of it before I asked the question. But that's not exactly a satisfying formal proof, in my opinion.
– James Shapiro
Nov 20 '18 at 1:24




















draft saved

draft discarded




















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.





Some of your past answers have not been well-received, and you're in danger of being blocked from answering.


Please pay close attention to the following guidance:


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3005793%2fsimple-proof-of-the-following-a-matrix-a-is-onto-if-and-only-if-its-rows-are%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

MongoDB - Not Authorized To Execute Command

How to fix TextFormField cause rebuild widget in Flutter

in spring boot 2.1 many test slices are not allowed anymore due to multiple @BootstrapWith