Optimize $M$ such that $MM^T$ is the “smallest” (in a positive semi-definite sense)











up vote
2
down vote

favorite
2












I want to solve the following problem:



begin{alignat}{}
underset{M}{text{minimize}} quad & MM^T \
text{subject to} quad & MF=I.
end{alignat}



where by minimize $MM^T$ I mean to find $M$ such that for any other matrix $hat M$ that satisfies the constraint, I have
$$
MM^T le hat Mhat M^T.
$$

that is, $MM^T - hat Mhat M^T$ is negative semi-definite. $F$ is (full rank) $n times p$ with $n>p$ and $I$ the identity matrix.



How can I translate this "smallest" concept into the objective function of the minimization problem? Should I perhaps minimize some kind of norm of $MM^T$? And after formalizing the problem, what is its solution?



Edit 1: For context, I am working on the minimum-variance unbiased linear estimator problem.



I am given the vector $y$ and a matrix $F$ such that



$$
y = Fx + epsilon
$$

where $F in mathbb{R}^{m times n}$ ($m>n$) full rank, $y in mathbb{R}^m$, $x in mathbb{R}^n$ and $epsilon$ is a random vector with zero mean and covariance matrix $I in mathbb{R}^{m times m}$ (identity matrix) and I want to estimate $x$.



I want to derive a linear estimator ($hat x = My$) such that it is unbiased ($E[hat x]=x$) and its covariance matrix satisfies
$$
E[(hat x - x)(hat x - x)^T] le E[(tilde{x} - x)(tilde{x} - x)^T]
$$

for any unbiased linear estimate $tilde{x}=tilde{M}y$.



For this problem, I know that such a linear estimator is unbiased if $MF=I$ (identity matrix) and its covariance matrix is
$$
E[(hat x - x)(hat x - x)^T] = MM^T
$$

and this is the point I am stuck.



I know that the solution for this problem is $M=(F^TF)^{-1}F^T$, and all the proofs I found assume this $M$ as an estimator, and then go on to prove that it is in fact the minimum-variance unbiased one. However, I want to arrive at this result starting from first principles, as if I did not know the solution a priori.



Edit 2: For reference, this pdf also discusses the topic using this informal notion of "smallest" and "largest" covariance matrix.



Edit 3: After some numerical tests, I got to the conclusion that minimizing $| MM^T |_2$ or $| MM^T |_{F}$ will both lead to the desired answer, that is, $M = (F^TF)^{-1}F^T$. Could someone give a formal argument for why this is the case?










share|cite|improve this question




















  • 1




    Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
    – Jean Marie
    2 days ago








  • 1




    @JeanMarie thanks, I edited the question.
    – pedroszattoni
    2 days ago






  • 2




    not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
    – LinAlg
    2 days ago








  • 1




    Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
    – user1551
    2 days ago








  • 1




    I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
    – Eric
    yesterday

















up vote
2
down vote

favorite
2












I want to solve the following problem:



begin{alignat}{}
underset{M}{text{minimize}} quad & MM^T \
text{subject to} quad & MF=I.
end{alignat}



where by minimize $MM^T$ I mean to find $M$ such that for any other matrix $hat M$ that satisfies the constraint, I have
$$
MM^T le hat Mhat M^T.
$$

that is, $MM^T - hat Mhat M^T$ is negative semi-definite. $F$ is (full rank) $n times p$ with $n>p$ and $I$ the identity matrix.



How can I translate this "smallest" concept into the objective function of the minimization problem? Should I perhaps minimize some kind of norm of $MM^T$? And after formalizing the problem, what is its solution?



Edit 1: For context, I am working on the minimum-variance unbiased linear estimator problem.



I am given the vector $y$ and a matrix $F$ such that



$$
y = Fx + epsilon
$$

where $F in mathbb{R}^{m times n}$ ($m>n$) full rank, $y in mathbb{R}^m$, $x in mathbb{R}^n$ and $epsilon$ is a random vector with zero mean and covariance matrix $I in mathbb{R}^{m times m}$ (identity matrix) and I want to estimate $x$.



I want to derive a linear estimator ($hat x = My$) such that it is unbiased ($E[hat x]=x$) and its covariance matrix satisfies
$$
E[(hat x - x)(hat x - x)^T] le E[(tilde{x} - x)(tilde{x} - x)^T]
$$

for any unbiased linear estimate $tilde{x}=tilde{M}y$.



For this problem, I know that such a linear estimator is unbiased if $MF=I$ (identity matrix) and its covariance matrix is
$$
E[(hat x - x)(hat x - x)^T] = MM^T
$$

and this is the point I am stuck.



I know that the solution for this problem is $M=(F^TF)^{-1}F^T$, and all the proofs I found assume this $M$ as an estimator, and then go on to prove that it is in fact the minimum-variance unbiased one. However, I want to arrive at this result starting from first principles, as if I did not know the solution a priori.



Edit 2: For reference, this pdf also discusses the topic using this informal notion of "smallest" and "largest" covariance matrix.



Edit 3: After some numerical tests, I got to the conclusion that minimizing $| MM^T |_2$ or $| MM^T |_{F}$ will both lead to the desired answer, that is, $M = (F^TF)^{-1}F^T$. Could someone give a formal argument for why this is the case?










share|cite|improve this question




















  • 1




    Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
    – Jean Marie
    2 days ago








  • 1




    @JeanMarie thanks, I edited the question.
    – pedroszattoni
    2 days ago






  • 2




    not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
    – LinAlg
    2 days ago








  • 1




    Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
    – user1551
    2 days ago








  • 1




    I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
    – Eric
    yesterday















up vote
2
down vote

favorite
2









up vote
2
down vote

favorite
2






2





I want to solve the following problem:



begin{alignat}{}
underset{M}{text{minimize}} quad & MM^T \
text{subject to} quad & MF=I.
end{alignat}



where by minimize $MM^T$ I mean to find $M$ such that for any other matrix $hat M$ that satisfies the constraint, I have
$$
MM^T le hat Mhat M^T.
$$

that is, $MM^T - hat Mhat M^T$ is negative semi-definite. $F$ is (full rank) $n times p$ with $n>p$ and $I$ the identity matrix.



How can I translate this "smallest" concept into the objective function of the minimization problem? Should I perhaps minimize some kind of norm of $MM^T$? And after formalizing the problem, what is its solution?



Edit 1: For context, I am working on the minimum-variance unbiased linear estimator problem.



I am given the vector $y$ and a matrix $F$ such that



$$
y = Fx + epsilon
$$

where $F in mathbb{R}^{m times n}$ ($m>n$) full rank, $y in mathbb{R}^m$, $x in mathbb{R}^n$ and $epsilon$ is a random vector with zero mean and covariance matrix $I in mathbb{R}^{m times m}$ (identity matrix) and I want to estimate $x$.



I want to derive a linear estimator ($hat x = My$) such that it is unbiased ($E[hat x]=x$) and its covariance matrix satisfies
$$
E[(hat x - x)(hat x - x)^T] le E[(tilde{x} - x)(tilde{x} - x)^T]
$$

for any unbiased linear estimate $tilde{x}=tilde{M}y$.



For this problem, I know that such a linear estimator is unbiased if $MF=I$ (identity matrix) and its covariance matrix is
$$
E[(hat x - x)(hat x - x)^T] = MM^T
$$

and this is the point I am stuck.



I know that the solution for this problem is $M=(F^TF)^{-1}F^T$, and all the proofs I found assume this $M$ as an estimator, and then go on to prove that it is in fact the minimum-variance unbiased one. However, I want to arrive at this result starting from first principles, as if I did not know the solution a priori.



Edit 2: For reference, this pdf also discusses the topic using this informal notion of "smallest" and "largest" covariance matrix.



Edit 3: After some numerical tests, I got to the conclusion that minimizing $| MM^T |_2$ or $| MM^T |_{F}$ will both lead to the desired answer, that is, $M = (F^TF)^{-1}F^T$. Could someone give a formal argument for why this is the case?










share|cite|improve this question















I want to solve the following problem:



begin{alignat}{}
underset{M}{text{minimize}} quad & MM^T \
text{subject to} quad & MF=I.
end{alignat}



where by minimize $MM^T$ I mean to find $M$ such that for any other matrix $hat M$ that satisfies the constraint, I have
$$
MM^T le hat Mhat M^T.
$$

that is, $MM^T - hat Mhat M^T$ is negative semi-definite. $F$ is (full rank) $n times p$ with $n>p$ and $I$ the identity matrix.



How can I translate this "smallest" concept into the objective function of the minimization problem? Should I perhaps minimize some kind of norm of $MM^T$? And after formalizing the problem, what is its solution?



Edit 1: For context, I am working on the minimum-variance unbiased linear estimator problem.



I am given the vector $y$ and a matrix $F$ such that



$$
y = Fx + epsilon
$$

where $F in mathbb{R}^{m times n}$ ($m>n$) full rank, $y in mathbb{R}^m$, $x in mathbb{R}^n$ and $epsilon$ is a random vector with zero mean and covariance matrix $I in mathbb{R}^{m times m}$ (identity matrix) and I want to estimate $x$.



I want to derive a linear estimator ($hat x = My$) such that it is unbiased ($E[hat x]=x$) and its covariance matrix satisfies
$$
E[(hat x - x)(hat x - x)^T] le E[(tilde{x} - x)(tilde{x} - x)^T]
$$

for any unbiased linear estimate $tilde{x}=tilde{M}y$.



For this problem, I know that such a linear estimator is unbiased if $MF=I$ (identity matrix) and its covariance matrix is
$$
E[(hat x - x)(hat x - x)^T] = MM^T
$$

and this is the point I am stuck.



I know that the solution for this problem is $M=(F^TF)^{-1}F^T$, and all the proofs I found assume this $M$ as an estimator, and then go on to prove that it is in fact the minimum-variance unbiased one. However, I want to arrive at this result starting from first principles, as if I did not know the solution a priori.



Edit 2: For reference, this pdf also discusses the topic using this informal notion of "smallest" and "largest" covariance matrix.



Edit 3: After some numerical tests, I got to the conclusion that minimizing $| MM^T |_2$ or $| MM^T |_{F}$ will both lead to the desired answer, that is, $M = (F^TF)^{-1}F^T$. Could someone give a formal argument for why this is the case?







linear-algebra optimization estimation least-squares covariance






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited yesterday

























asked 2 days ago









pedroszattoni

183110




183110








  • 1




    Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
    – Jean Marie
    2 days ago








  • 1




    @JeanMarie thanks, I edited the question.
    – pedroszattoni
    2 days ago






  • 2




    not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
    – LinAlg
    2 days ago








  • 1




    Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
    – user1551
    2 days ago








  • 1




    I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
    – Eric
    yesterday
















  • 1




    Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
    – Jean Marie
    2 days ago








  • 1




    @JeanMarie thanks, I edited the question.
    – pedroszattoni
    2 days ago






  • 2




    not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
    – LinAlg
    2 days ago








  • 1




    Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
    – user1551
    2 days ago








  • 1




    I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
    – Eric
    yesterday










1




1




Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
– Jean Marie
2 days ago






Ah... "skinny" and "fat" are rather uncommon denominations. Prefer for example "$F$ is $n times p$ with $n>p$"
– Jean Marie
2 days ago






1




1




@JeanMarie thanks, I edited the question.
– pedroszattoni
2 days ago




@JeanMarie thanks, I edited the question.
– pedroszattoni
2 days ago




2




2




not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
– LinAlg
2 days ago






not all pd matrices can be compared, e.g., when A=diag(2,1) and B=diag(1,2), then neither $Aleq B$ nor $Ageq B$.
– LinAlg
2 days ago






1




1




Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
– user1551
2 days ago






Your question is still unclear. You should clarify the meaning of $Ale B$ for two matrices $A$ and $B$. Your question title seems to suggest that you are talking about positive semidefinite partial ordering, but in the question body, when you are comparing covariance matrices, my impression is that you are talking about entrywise comparison.
– user1551
2 days ago






1




1




I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
– Eric
yesterday






I seem to recall that every left inverse of a matrix $F$ can be expressed in the form $(F^* F)^{-1} F^*+K$ where $K$ is a matrix which satisfies $KF=0$. Once you have this, it is not too hard to show that there is a smallest solution (in the ordering $A leq B$ means $B-A$ is positive semidefinite) and that it is obtained by taking $K=0$. I am leaving this as a comment rather than an answer since I don't fully remember on the claim that every left inverse has this form.
– Eric
yesterday

















active

oldest

votes











Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














 

draft saved


draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3005533%2foptimize-m-such-that-mmt-is-the-smallest-in-a-positive-semi-definite-sen%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown






























active

oldest

votes













active

oldest

votes









active

oldest

votes






active

oldest

votes
















 

draft saved


draft discarded



















































 


draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3005533%2foptimize-m-such-that-mmt-is-the-smallest-in-a-positive-semi-definite-sen%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

MongoDB - Not Authorized To Execute Command

Npm cannot find a required file even through it is in the searched directory

in spring boot 2.1 many test slices are not allowed anymore due to multiple @BootstrapWith