Covariance of polynomials of random normal variables
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$If $X$ and $Y$ are random variables with a bivariate normal distribution and:
- $Xsimmathcal{N}(mu_X,sigma_X^2)$
- $Ysimmathcal{N}(mu_Y,sigma_Y^2)$
- $Cov(X,Y)neq0$
May I compute $Cov(X^m,Y^n)$ for arbitrary positive integers $m$ and $n$?
probability statistics random-variables normal-distribution covariance
$endgroup$
|
show 2 more comments
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$If $X$ and $Y$ are random variables with a bivariate normal distribution and:
- $Xsimmathcal{N}(mu_X,sigma_X^2)$
- $Ysimmathcal{N}(mu_Y,sigma_Y^2)$
- $Cov(X,Y)neq0$
May I compute $Cov(X^m,Y^n)$ for arbitrary positive integers $m$ and $n$?
probability statistics random-variables normal-distribution covariance
$endgroup$
1
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44
|
show 2 more comments
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$If $X$ and $Y$ are random variables with a bivariate normal distribution and:
- $Xsimmathcal{N}(mu_X,sigma_X^2)$
- $Ysimmathcal{N}(mu_Y,sigma_Y^2)$
- $Cov(X,Y)neq0$
May I compute $Cov(X^m,Y^n)$ for arbitrary positive integers $m$ and $n$?
probability statistics random-variables normal-distribution covariance
$endgroup$
$newcommand{Cov}{operatorname{Cov}}$If $X$ and $Y$ are random variables with a bivariate normal distribution and:
- $Xsimmathcal{N}(mu_X,sigma_X^2)$
- $Ysimmathcal{N}(mu_Y,sigma_Y^2)$
- $Cov(X,Y)neq0$
May I compute $Cov(X^m,Y^n)$ for arbitrary positive integers $m$ and $n$?
probability statistics random-variables normal-distribution covariance
probability statistics random-variables normal-distribution covariance
edited Aug 8 '18 at 6:03
Mathieu
asked Aug 7 '18 at 22:13
MathieuMathieu
1135
1135
1
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44
|
show 2 more comments
1
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44
1
1
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44
|
show 2 more comments
3 Answers
3
active
oldest
votes
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$Given the new background that the OP provided to his question (most general normal bivariate distribution for variables (X_1, X_2) with non-zero covariance), it is possible to find a general formula for $Cov(X_1^m, X_2^n)$ as follows:
Consider the generating function $phi(t_1,t_2)$ for the bivariate distribution given here, in equation (57) (a clean derivation is given for it so I won't repeat it). Then:
$$E(X_1^m X_2^n)=(-i)^{m+n}(frac{partial}{partial t_1})^m(frac{partial}{partial t_2})^nphi(t_1,t_2)Big|_{(t_1, t_2)=(0,0)} tag 1$$
After some tedious and careful algebra we wish to write the generating function $phi(t_1,t_2)=e^{it_1mu_1+it_2mu_2}e^{-frac{1}{2}(sigma_1^2t_1^2+2rhosigma_1t_1sigma_2t_2+sigma_2^2t_2^2)}$ in the form:
$$phi(t_1,t_2)=e^{-a}e^{-b(t_2+c)^2}e^{-d(t_1+gt_2+h)^2}tag2$$
which is possible for the values:
$$a=frac{mu_1^2}{2sigma_1^2}+frac{(mu_2-mu_1rho frac{sigma_2}{sigma_1})^2}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}b=frac{1}{2}(1-rho^2)sigma_{2}^2, hspace{0.2cm}c=-ifrac{mu_2-mu_1rho frac{sigma_2}{sigma_1}}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}d=frac{sigma_1^2}{2},hspace{0.2cm} g=frac{rhosigma_2}{sigma_1},hspace{0.2cm}h=-ifrac{mu_1}{sigma_1^2} $$
We change variables
$y_2=sqrt{b}(t_2+c), y_1=sqrt{d}(t_1+gt_2+h)$ and we find:
$$begin{align}frac{partial}{partial t_1}&=sqrt{d}frac{partial}{partial y_1}\ frac{partial}{partial t_2}&=gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2}end{align}$$
Substitute into (1) for the result:
$$E(X_1^mX_2^n)=(-i)^{m+n}(sqrt{d}frac{partial}{partial y_1})^m(gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2})^n e^{-a}e^{-y_1^2}e^{-y^2_2}Big|_{(y_1, y_2)=(csqrt{b},hsqrt{d})}$$
Expanding the parentheses and using the Rodrigues formula for Hermite polynomials we get:
$$E(X_1^mX_2^n)=e^{-a-bc^2-dh^2}i^{m+n}g^n d^{frac{m+n}{2}}sum_{k=0}^nfrac{n!}{k!(n-k)!}Big(frac{sqrt{b}}{gsqrt{d}}Big)^{n-k}H_{n-k}(hsqrt{d})H_{m+k}(csqrt{b})$$
The expectation values $E(X_1^m)$ and $E(X_2^n)$ can be calculated by setting $n=0$, $m=0$ in the general formula respectively.
$endgroup$
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
add a comment |
$begingroup$
I do not think that there is a nice answer here, sorry.$newcommand{Cov}{operatorname{Cov}}$
Note firstly that
$$
Cov(X^m,Y^n)=E(X^mY^n)-E(X^m)E(Y^n)
$$
by definition of the covariance.
This answer provides you with:
$$E(X^m)=sum_{k=0}^{lfloor m/2rfloor} {m choose 2k}(2k-1)!!sigma_X^{2k}mu_X^{m-2k}.$$
and similar for $Y^n$. However, evaluating $E(X^mY^n)$ is the issue. See here how messy the distribution of just two non-independent normal random variables is, yet you have the product of $m+n$ of them!
$endgroup$
add a comment |
$begingroup$
You have said$newcommand{Cov}{operatorname{Cov}}$
begin{align}
& Xsim N(mu_X,sigma^2_X), \[4pt]
& Y sim N(mu_Y,sigma^2_Y), \[4pt]
& Cov(X,Y) = 0.
end{align}
That falls short of specifying the joint distribution of $(X,Y).$ If it were further specified that $X,Y$ are jointly normally distributed, then the covariance can be $0$ only if $X,Y$ are independent. If $X,Y$ are independent, then so are $X^m,Y^n,$ so their covariance is also $0.$
Here is a simple example: Suppose $Xsim N(mu_X, sigma_X^2),$ let $Z = text{the “z-score''} = (X-mu_X)/sigma_X,$ and independently of $X$ you toss a coin. Then let $Y = mu_Y pm sigma_Y Z, $ where the choice between $text{“}pmtext{''}$ is determined by the coin toss. Then $X,Y$ have covariance $0$ and have just the distributions you specified in the question, but they are NOT JOINTLY normally distributed and not independent.
But if you assume joint normality, which in the case of two random variables means bivariate normality, then the answer is just as in the first paragraph above.
$endgroup$
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
|
show 1 more comment
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2875466%2fcovariance-of-polynomials-of-random-normal-variables%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$Given the new background that the OP provided to his question (most general normal bivariate distribution for variables (X_1, X_2) with non-zero covariance), it is possible to find a general formula for $Cov(X_1^m, X_2^n)$ as follows:
Consider the generating function $phi(t_1,t_2)$ for the bivariate distribution given here, in equation (57) (a clean derivation is given for it so I won't repeat it). Then:
$$E(X_1^m X_2^n)=(-i)^{m+n}(frac{partial}{partial t_1})^m(frac{partial}{partial t_2})^nphi(t_1,t_2)Big|_{(t_1, t_2)=(0,0)} tag 1$$
After some tedious and careful algebra we wish to write the generating function $phi(t_1,t_2)=e^{it_1mu_1+it_2mu_2}e^{-frac{1}{2}(sigma_1^2t_1^2+2rhosigma_1t_1sigma_2t_2+sigma_2^2t_2^2)}$ in the form:
$$phi(t_1,t_2)=e^{-a}e^{-b(t_2+c)^2}e^{-d(t_1+gt_2+h)^2}tag2$$
which is possible for the values:
$$a=frac{mu_1^2}{2sigma_1^2}+frac{(mu_2-mu_1rho frac{sigma_2}{sigma_1})^2}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}b=frac{1}{2}(1-rho^2)sigma_{2}^2, hspace{0.2cm}c=-ifrac{mu_2-mu_1rho frac{sigma_2}{sigma_1}}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}d=frac{sigma_1^2}{2},hspace{0.2cm} g=frac{rhosigma_2}{sigma_1},hspace{0.2cm}h=-ifrac{mu_1}{sigma_1^2} $$
We change variables
$y_2=sqrt{b}(t_2+c), y_1=sqrt{d}(t_1+gt_2+h)$ and we find:
$$begin{align}frac{partial}{partial t_1}&=sqrt{d}frac{partial}{partial y_1}\ frac{partial}{partial t_2}&=gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2}end{align}$$
Substitute into (1) for the result:
$$E(X_1^mX_2^n)=(-i)^{m+n}(sqrt{d}frac{partial}{partial y_1})^m(gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2})^n e^{-a}e^{-y_1^2}e^{-y^2_2}Big|_{(y_1, y_2)=(csqrt{b},hsqrt{d})}$$
Expanding the parentheses and using the Rodrigues formula for Hermite polynomials we get:
$$E(X_1^mX_2^n)=e^{-a-bc^2-dh^2}i^{m+n}g^n d^{frac{m+n}{2}}sum_{k=0}^nfrac{n!}{k!(n-k)!}Big(frac{sqrt{b}}{gsqrt{d}}Big)^{n-k}H_{n-k}(hsqrt{d})H_{m+k}(csqrt{b})$$
The expectation values $E(X_1^m)$ and $E(X_2^n)$ can be calculated by setting $n=0$, $m=0$ in the general formula respectively.
$endgroup$
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
add a comment |
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$Given the new background that the OP provided to his question (most general normal bivariate distribution for variables (X_1, X_2) with non-zero covariance), it is possible to find a general formula for $Cov(X_1^m, X_2^n)$ as follows:
Consider the generating function $phi(t_1,t_2)$ for the bivariate distribution given here, in equation (57) (a clean derivation is given for it so I won't repeat it). Then:
$$E(X_1^m X_2^n)=(-i)^{m+n}(frac{partial}{partial t_1})^m(frac{partial}{partial t_2})^nphi(t_1,t_2)Big|_{(t_1, t_2)=(0,0)} tag 1$$
After some tedious and careful algebra we wish to write the generating function $phi(t_1,t_2)=e^{it_1mu_1+it_2mu_2}e^{-frac{1}{2}(sigma_1^2t_1^2+2rhosigma_1t_1sigma_2t_2+sigma_2^2t_2^2)}$ in the form:
$$phi(t_1,t_2)=e^{-a}e^{-b(t_2+c)^2}e^{-d(t_1+gt_2+h)^2}tag2$$
which is possible for the values:
$$a=frac{mu_1^2}{2sigma_1^2}+frac{(mu_2-mu_1rho frac{sigma_2}{sigma_1})^2}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}b=frac{1}{2}(1-rho^2)sigma_{2}^2, hspace{0.2cm}c=-ifrac{mu_2-mu_1rho frac{sigma_2}{sigma_1}}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}d=frac{sigma_1^2}{2},hspace{0.2cm} g=frac{rhosigma_2}{sigma_1},hspace{0.2cm}h=-ifrac{mu_1}{sigma_1^2} $$
We change variables
$y_2=sqrt{b}(t_2+c), y_1=sqrt{d}(t_1+gt_2+h)$ and we find:
$$begin{align}frac{partial}{partial t_1}&=sqrt{d}frac{partial}{partial y_1}\ frac{partial}{partial t_2}&=gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2}end{align}$$
Substitute into (1) for the result:
$$E(X_1^mX_2^n)=(-i)^{m+n}(sqrt{d}frac{partial}{partial y_1})^m(gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2})^n e^{-a}e^{-y_1^2}e^{-y^2_2}Big|_{(y_1, y_2)=(csqrt{b},hsqrt{d})}$$
Expanding the parentheses and using the Rodrigues formula for Hermite polynomials we get:
$$E(X_1^mX_2^n)=e^{-a-bc^2-dh^2}i^{m+n}g^n d^{frac{m+n}{2}}sum_{k=0}^nfrac{n!}{k!(n-k)!}Big(frac{sqrt{b}}{gsqrt{d}}Big)^{n-k}H_{n-k}(hsqrt{d})H_{m+k}(csqrt{b})$$
The expectation values $E(X_1^m)$ and $E(X_2^n)$ can be calculated by setting $n=0$, $m=0$ in the general formula respectively.
$endgroup$
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
add a comment |
$begingroup$
$newcommand{Cov}{operatorname{Cov}}$Given the new background that the OP provided to his question (most general normal bivariate distribution for variables (X_1, X_2) with non-zero covariance), it is possible to find a general formula for $Cov(X_1^m, X_2^n)$ as follows:
Consider the generating function $phi(t_1,t_2)$ for the bivariate distribution given here, in equation (57) (a clean derivation is given for it so I won't repeat it). Then:
$$E(X_1^m X_2^n)=(-i)^{m+n}(frac{partial}{partial t_1})^m(frac{partial}{partial t_2})^nphi(t_1,t_2)Big|_{(t_1, t_2)=(0,0)} tag 1$$
After some tedious and careful algebra we wish to write the generating function $phi(t_1,t_2)=e^{it_1mu_1+it_2mu_2}e^{-frac{1}{2}(sigma_1^2t_1^2+2rhosigma_1t_1sigma_2t_2+sigma_2^2t_2^2)}$ in the form:
$$phi(t_1,t_2)=e^{-a}e^{-b(t_2+c)^2}e^{-d(t_1+gt_2+h)^2}tag2$$
which is possible for the values:
$$a=frac{mu_1^2}{2sigma_1^2}+frac{(mu_2-mu_1rho frac{sigma_2}{sigma_1})^2}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}b=frac{1}{2}(1-rho^2)sigma_{2}^2, hspace{0.2cm}c=-ifrac{mu_2-mu_1rho frac{sigma_2}{sigma_1}}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}d=frac{sigma_1^2}{2},hspace{0.2cm} g=frac{rhosigma_2}{sigma_1},hspace{0.2cm}h=-ifrac{mu_1}{sigma_1^2} $$
We change variables
$y_2=sqrt{b}(t_2+c), y_1=sqrt{d}(t_1+gt_2+h)$ and we find:
$$begin{align}frac{partial}{partial t_1}&=sqrt{d}frac{partial}{partial y_1}\ frac{partial}{partial t_2}&=gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2}end{align}$$
Substitute into (1) for the result:
$$E(X_1^mX_2^n)=(-i)^{m+n}(sqrt{d}frac{partial}{partial y_1})^m(gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2})^n e^{-a}e^{-y_1^2}e^{-y^2_2}Big|_{(y_1, y_2)=(csqrt{b},hsqrt{d})}$$
Expanding the parentheses and using the Rodrigues formula for Hermite polynomials we get:
$$E(X_1^mX_2^n)=e^{-a-bc^2-dh^2}i^{m+n}g^n d^{frac{m+n}{2}}sum_{k=0}^nfrac{n!}{k!(n-k)!}Big(frac{sqrt{b}}{gsqrt{d}}Big)^{n-k}H_{n-k}(hsqrt{d})H_{m+k}(csqrt{b})$$
The expectation values $E(X_1^m)$ and $E(X_2^n)$ can be calculated by setting $n=0$, $m=0$ in the general formula respectively.
$endgroup$
$newcommand{Cov}{operatorname{Cov}}$Given the new background that the OP provided to his question (most general normal bivariate distribution for variables (X_1, X_2) with non-zero covariance), it is possible to find a general formula for $Cov(X_1^m, X_2^n)$ as follows:
Consider the generating function $phi(t_1,t_2)$ for the bivariate distribution given here, in equation (57) (a clean derivation is given for it so I won't repeat it). Then:
$$E(X_1^m X_2^n)=(-i)^{m+n}(frac{partial}{partial t_1})^m(frac{partial}{partial t_2})^nphi(t_1,t_2)Big|_{(t_1, t_2)=(0,0)} tag 1$$
After some tedious and careful algebra we wish to write the generating function $phi(t_1,t_2)=e^{it_1mu_1+it_2mu_2}e^{-frac{1}{2}(sigma_1^2t_1^2+2rhosigma_1t_1sigma_2t_2+sigma_2^2t_2^2)}$ in the form:
$$phi(t_1,t_2)=e^{-a}e^{-b(t_2+c)^2}e^{-d(t_1+gt_2+h)^2}tag2$$
which is possible for the values:
$$a=frac{mu_1^2}{2sigma_1^2}+frac{(mu_2-mu_1rho frac{sigma_2}{sigma_1})^2}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}b=frac{1}{2}(1-rho^2)sigma_{2}^2, hspace{0.2cm}c=-ifrac{mu_2-mu_1rho frac{sigma_2}{sigma_1}}{2sigma^2_2(1-rho^2)}, hspace{0.2cm}d=frac{sigma_1^2}{2},hspace{0.2cm} g=frac{rhosigma_2}{sigma_1},hspace{0.2cm}h=-ifrac{mu_1}{sigma_1^2} $$
We change variables
$y_2=sqrt{b}(t_2+c), y_1=sqrt{d}(t_1+gt_2+h)$ and we find:
$$begin{align}frac{partial}{partial t_1}&=sqrt{d}frac{partial}{partial y_1}\ frac{partial}{partial t_2}&=gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2}end{align}$$
Substitute into (1) for the result:
$$E(X_1^mX_2^n)=(-i)^{m+n}(sqrt{d}frac{partial}{partial y_1})^m(gsqrt{d}frac{partial}{partial y_1}+sqrt{b}frac{partial}{partial y_2})^n e^{-a}e^{-y_1^2}e^{-y^2_2}Big|_{(y_1, y_2)=(csqrt{b},hsqrt{d})}$$
Expanding the parentheses and using the Rodrigues formula for Hermite polynomials we get:
$$E(X_1^mX_2^n)=e^{-a-bc^2-dh^2}i^{m+n}g^n d^{frac{m+n}{2}}sum_{k=0}^nfrac{n!}{k!(n-k)!}Big(frac{sqrt{b}}{gsqrt{d}}Big)^{n-k}H_{n-k}(hsqrt{d})H_{m+k}(csqrt{b})$$
The expectation values $E(X_1^m)$ and $E(X_2^n)$ can be calculated by setting $n=0$, $m=0$ in the general formula respectively.
edited Jan 29 at 11:39
Martin Sleziak
44.9k10122277
44.9k10122277
answered Aug 8 '18 at 23:01
DinosaurEggDinosaurEgg
49017
49017
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
add a comment |
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
This makes the problem far more complicated than it is, unless you don't know that JOINT normality plus uncorrelatedness entails independence.
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:08
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
The OP has stated that the two variables are correlated since their covariance is non-zero. The OP also updated their request with a joint probability that reflects this. This problem is valid and I answered with a calculation that addresses it. The formula is simple and a finite polynomial in all variables except for $rho$ so I dont see how this problem is complicated.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:43
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
$begingroup$
I don't think the OP gave a good description of what the problem was in it's initial wording, but then they changed their mind. It seems to me that to them quadratic correlations are more important than solving the totally trivial independent variable problem.
$endgroup$
– DinosaurEgg
Aug 9 '18 at 3:48
add a comment |
$begingroup$
I do not think that there is a nice answer here, sorry.$newcommand{Cov}{operatorname{Cov}}$
Note firstly that
$$
Cov(X^m,Y^n)=E(X^mY^n)-E(X^m)E(Y^n)
$$
by definition of the covariance.
This answer provides you with:
$$E(X^m)=sum_{k=0}^{lfloor m/2rfloor} {m choose 2k}(2k-1)!!sigma_X^{2k}mu_X^{m-2k}.$$
and similar for $Y^n$. However, evaluating $E(X^mY^n)$ is the issue. See here how messy the distribution of just two non-independent normal random variables is, yet you have the product of $m+n$ of them!
$endgroup$
add a comment |
$begingroup$
I do not think that there is a nice answer here, sorry.$newcommand{Cov}{operatorname{Cov}}$
Note firstly that
$$
Cov(X^m,Y^n)=E(X^mY^n)-E(X^m)E(Y^n)
$$
by definition of the covariance.
This answer provides you with:
$$E(X^m)=sum_{k=0}^{lfloor m/2rfloor} {m choose 2k}(2k-1)!!sigma_X^{2k}mu_X^{m-2k}.$$
and similar for $Y^n$. However, evaluating $E(X^mY^n)$ is the issue. See here how messy the distribution of just two non-independent normal random variables is, yet you have the product of $m+n$ of them!
$endgroup$
add a comment |
$begingroup$
I do not think that there is a nice answer here, sorry.$newcommand{Cov}{operatorname{Cov}}$
Note firstly that
$$
Cov(X^m,Y^n)=E(X^mY^n)-E(X^m)E(Y^n)
$$
by definition of the covariance.
This answer provides you with:
$$E(X^m)=sum_{k=0}^{lfloor m/2rfloor} {m choose 2k}(2k-1)!!sigma_X^{2k}mu_X^{m-2k}.$$
and similar for $Y^n$. However, evaluating $E(X^mY^n)$ is the issue. See here how messy the distribution of just two non-independent normal random variables is, yet you have the product of $m+n$ of them!
$endgroup$
I do not think that there is a nice answer here, sorry.$newcommand{Cov}{operatorname{Cov}}$
Note firstly that
$$
Cov(X^m,Y^n)=E(X^mY^n)-E(X^m)E(Y^n)
$$
by definition of the covariance.
This answer provides you with:
$$E(X^m)=sum_{k=0}^{lfloor m/2rfloor} {m choose 2k}(2k-1)!!sigma_X^{2k}mu_X^{m-2k}.$$
and similar for $Y^n$. However, evaluating $E(X^mY^n)$ is the issue. See here how messy the distribution of just two non-independent normal random variables is, yet you have the product of $m+n$ of them!
edited Jan 29 at 11:39
Martin Sleziak
44.9k10122277
44.9k10122277
answered Aug 7 '18 at 22:37
MalkinMalkin
1,512726
1,512726
add a comment |
add a comment |
$begingroup$
You have said$newcommand{Cov}{operatorname{Cov}}$
begin{align}
& Xsim N(mu_X,sigma^2_X), \[4pt]
& Y sim N(mu_Y,sigma^2_Y), \[4pt]
& Cov(X,Y) = 0.
end{align}
That falls short of specifying the joint distribution of $(X,Y).$ If it were further specified that $X,Y$ are jointly normally distributed, then the covariance can be $0$ only if $X,Y$ are independent. If $X,Y$ are independent, then so are $X^m,Y^n,$ so their covariance is also $0.$
Here is a simple example: Suppose $Xsim N(mu_X, sigma_X^2),$ let $Z = text{the “z-score''} = (X-mu_X)/sigma_X,$ and independently of $X$ you toss a coin. Then let $Y = mu_Y pm sigma_Y Z, $ where the choice between $text{“}pmtext{''}$ is determined by the coin toss. Then $X,Y$ have covariance $0$ and have just the distributions you specified in the question, but they are NOT JOINTLY normally distributed and not independent.
But if you assume joint normality, which in the case of two random variables means bivariate normality, then the answer is just as in the first paragraph above.
$endgroup$
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
|
show 1 more comment
$begingroup$
You have said$newcommand{Cov}{operatorname{Cov}}$
begin{align}
& Xsim N(mu_X,sigma^2_X), \[4pt]
& Y sim N(mu_Y,sigma^2_Y), \[4pt]
& Cov(X,Y) = 0.
end{align}
That falls short of specifying the joint distribution of $(X,Y).$ If it were further specified that $X,Y$ are jointly normally distributed, then the covariance can be $0$ only if $X,Y$ are independent. If $X,Y$ are independent, then so are $X^m,Y^n,$ so their covariance is also $0.$
Here is a simple example: Suppose $Xsim N(mu_X, sigma_X^2),$ let $Z = text{the “z-score''} = (X-mu_X)/sigma_X,$ and independently of $X$ you toss a coin. Then let $Y = mu_Y pm sigma_Y Z, $ where the choice between $text{“}pmtext{''}$ is determined by the coin toss. Then $X,Y$ have covariance $0$ and have just the distributions you specified in the question, but they are NOT JOINTLY normally distributed and not independent.
But if you assume joint normality, which in the case of two random variables means bivariate normality, then the answer is just as in the first paragraph above.
$endgroup$
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
|
show 1 more comment
$begingroup$
You have said$newcommand{Cov}{operatorname{Cov}}$
begin{align}
& Xsim N(mu_X,sigma^2_X), \[4pt]
& Y sim N(mu_Y,sigma^2_Y), \[4pt]
& Cov(X,Y) = 0.
end{align}
That falls short of specifying the joint distribution of $(X,Y).$ If it were further specified that $X,Y$ are jointly normally distributed, then the covariance can be $0$ only if $X,Y$ are independent. If $X,Y$ are independent, then so are $X^m,Y^n,$ so their covariance is also $0.$
Here is a simple example: Suppose $Xsim N(mu_X, sigma_X^2),$ let $Z = text{the “z-score''} = (X-mu_X)/sigma_X,$ and independently of $X$ you toss a coin. Then let $Y = mu_Y pm sigma_Y Z, $ where the choice between $text{“}pmtext{''}$ is determined by the coin toss. Then $X,Y$ have covariance $0$ and have just the distributions you specified in the question, but they are NOT JOINTLY normally distributed and not independent.
But if you assume joint normality, which in the case of two random variables means bivariate normality, then the answer is just as in the first paragraph above.
$endgroup$
You have said$newcommand{Cov}{operatorname{Cov}}$
begin{align}
& Xsim N(mu_X,sigma^2_X), \[4pt]
& Y sim N(mu_Y,sigma^2_Y), \[4pt]
& Cov(X,Y) = 0.
end{align}
That falls short of specifying the joint distribution of $(X,Y).$ If it were further specified that $X,Y$ are jointly normally distributed, then the covariance can be $0$ only if $X,Y$ are independent. If $X,Y$ are independent, then so are $X^m,Y^n,$ so their covariance is also $0.$
Here is a simple example: Suppose $Xsim N(mu_X, sigma_X^2),$ let $Z = text{the “z-score''} = (X-mu_X)/sigma_X,$ and independently of $X$ you toss a coin. Then let $Y = mu_Y pm sigma_Y Z, $ where the choice between $text{“}pmtext{''}$ is determined by the coin toss. Then $X,Y$ have covariance $0$ and have just the distributions you specified in the question, but they are NOT JOINTLY normally distributed and not independent.
But if you assume joint normality, which in the case of two random variables means bivariate normality, then the answer is just as in the first paragraph above.
edited Jan 29 at 11:39
Martin Sleziak
44.9k10122277
44.9k10122277
answered Aug 8 '18 at 0:41
Michael HardyMichael Hardy
1
1
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
|
show 1 more comment
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
Thanks, but my original question states that Cov(X,Y) is not equal to zero. I edited it to specify a joint normal distribution.
$endgroup$
– Mathieu
Aug 8 '18 at 6:05
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
@Mathieu : In that case, my first paragraph above answers your question, and the answer by E.Malkin misses the point at best. There is no need to know the variance of $X^m$ or even the expected value of $X^m$ in order to answer your question.
$endgroup$
– Michael Hardy
Aug 8 '18 at 12:41
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
Sorry but I must be missing something: my question states that Cov(X,Y) is not zero and your first paragraph above states the opposite. Not sure what to make of that.
$endgroup$
– Mathieu
Aug 8 '18 at 13:14
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@MichaelHardy I also don't see how your first paragraph answers the question and could you please explain how I've missed the point ("at best")? Knowing $E(X^m)$ means that finding $Cov (X^m,Y^n)$ is equivalent to finding $E(X^mY^n)$, which surely isn't invalid?
$endgroup$
– Malkin
Aug 8 '18 at 22:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
$begingroup$
@Malkin : If $X,Y$ are not just separately normal but jointly normal, then having zero covariance entails that they are independent. If $X,Y$ are independent, then $X^m,Y^n$ are independent. If $X^m, Y^n$ are independent, then $operatorname E(X^m Y^n) = operatorname E(X^m) operatorname E(Y^n).$ Therefore $$operatorname{cov}(X^m, Y^n) = operatorname E(X^m Y^n) - operatorname E(X^m) operatorname E(Y^n) = operatorname E(X^m) operatorname E(Y^n) - operatorname E(X^m) operatorname E(Y^m) = 0.$$
$endgroup$
– Michael Hardy
Aug 9 '18 at 2:04
|
show 1 more comment
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2875466%2fcovariance-of-polynomials-of-random-normal-variables%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
1
$begingroup$
I don't think you can compute this object for arbitrary m and n. You would need to know more information about the random variables X, Y in the form of their joint distribution. Even if you assume that the covariance is well known, you still can't find the higher moments of the distribution.
$endgroup$
– DinosaurEgg
Aug 7 '18 at 22:32
$begingroup$
If I assume instead that $(X,Y)$ have a bivariate normal distribution whose variance-covariance matrix is known, does the question become answerable?
$endgroup$
– Mathieu
Aug 7 '18 at 22:52
$begingroup$
This question fails to state that $X,Y$ are JOINTLY normal. If that is assumed, then, since their covariance is $0,$ they are independent.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:32
$begingroup$
It is sloppy notation to write $Xsim N(mu_x, sigma^2_x)$ instead of $Xsim N(mu_X, sigma^2_X).$ And in many contexts when working with only slightly more involved problems of this kind, this sort of confusion can paralyze you.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:37
$begingroup$
In your comment you say $X,Y$ have a bivariate normal distribution, but in your question you say only that each one separately is normally distributed and their covariance is $0.$ I can show you examples of a distribution of a pair $(X,Y)$ in which each is normally distributed and their covariance is $0$ and they are NOT JOINTLY normally distributed. But if we assume bivariate (thus joint) normality, then the answer is easy. See my answer below.
$endgroup$
– Michael Hardy
Aug 8 '18 at 0:44