Intuition for the vanishing expectation of score? $E[S(theta)] =0$?












1












$begingroup$


Suppose $f$ is a probability density on $y in mathbb{R}$ and $theta in mathbb{R}$ is some parameter.



Define $$S(theta) := log big( frac{partial f(y; theta)}{partial theta} big)$$



Then I'm trying to show that $E[S(theta)] = 0$. The derivation shown in my class is first noticing that, when $L(theta; y) = f(y; theta)$ and then seeing that:



begin{align}
0 &= frac{partial 1}{partial theta}dy \
&= frac{partial}{partial theta} int f(y; theta) dy \
&= int f(y; theta) frac{ frac{ partial L(theta; y) }{partial theta} }{ L(theta; y) } dy \
&=int f(y; theta) frac{ partial log(L(theta; y)) }{ partial theta} dy\
&= E[S(theta)]
end{align}



I think that my lack of intuition for this derivation may come from a weak background in analysis, that is, I'm not so intuitively clear on what conditions are required for interchanging differentiation and integration and so perhaps without this intuition I don't see why this result is obvious.



So, I was trying to gain more intuition by proceeding more directly, starting with $E[S(theta)]$ and then unpacking the result until I arrived at $$E[S(theta)] = int frac{ partial f(y; theta) }{partial theta } dy$$



and from here, it is not obvious to me that I should be able to interchange integration and differentiation and so I don't know how to proceed directly.



Thanks in advance.










share|cite|improve this question









$endgroup$












  • $begingroup$
    The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
    $endgroup$
    – StubbornAtom
    Jan 25 at 18:48










  • $begingroup$
    Sure, but why do we need this assumption?
    $endgroup$
    – libby
    Jan 25 at 20:05
















1












$begingroup$


Suppose $f$ is a probability density on $y in mathbb{R}$ and $theta in mathbb{R}$ is some parameter.



Define $$S(theta) := log big( frac{partial f(y; theta)}{partial theta} big)$$



Then I'm trying to show that $E[S(theta)] = 0$. The derivation shown in my class is first noticing that, when $L(theta; y) = f(y; theta)$ and then seeing that:



begin{align}
0 &= frac{partial 1}{partial theta}dy \
&= frac{partial}{partial theta} int f(y; theta) dy \
&= int f(y; theta) frac{ frac{ partial L(theta; y) }{partial theta} }{ L(theta; y) } dy \
&=int f(y; theta) frac{ partial log(L(theta; y)) }{ partial theta} dy\
&= E[S(theta)]
end{align}



I think that my lack of intuition for this derivation may come from a weak background in analysis, that is, I'm not so intuitively clear on what conditions are required for interchanging differentiation and integration and so perhaps without this intuition I don't see why this result is obvious.



So, I was trying to gain more intuition by proceeding more directly, starting with $E[S(theta)]$ and then unpacking the result until I arrived at $$E[S(theta)] = int frac{ partial f(y; theta) }{partial theta } dy$$



and from here, it is not obvious to me that I should be able to interchange integration and differentiation and so I don't know how to proceed directly.



Thanks in advance.










share|cite|improve this question









$endgroup$












  • $begingroup$
    The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
    $endgroup$
    – StubbornAtom
    Jan 25 at 18:48










  • $begingroup$
    Sure, but why do we need this assumption?
    $endgroup$
    – libby
    Jan 25 at 20:05














1












1








1


1



$begingroup$


Suppose $f$ is a probability density on $y in mathbb{R}$ and $theta in mathbb{R}$ is some parameter.



Define $$S(theta) := log big( frac{partial f(y; theta)}{partial theta} big)$$



Then I'm trying to show that $E[S(theta)] = 0$. The derivation shown in my class is first noticing that, when $L(theta; y) = f(y; theta)$ and then seeing that:



begin{align}
0 &= frac{partial 1}{partial theta}dy \
&= frac{partial}{partial theta} int f(y; theta) dy \
&= int f(y; theta) frac{ frac{ partial L(theta; y) }{partial theta} }{ L(theta; y) } dy \
&=int f(y; theta) frac{ partial log(L(theta; y)) }{ partial theta} dy\
&= E[S(theta)]
end{align}



I think that my lack of intuition for this derivation may come from a weak background in analysis, that is, I'm not so intuitively clear on what conditions are required for interchanging differentiation and integration and so perhaps without this intuition I don't see why this result is obvious.



So, I was trying to gain more intuition by proceeding more directly, starting with $E[S(theta)]$ and then unpacking the result until I arrived at $$E[S(theta)] = int frac{ partial f(y; theta) }{partial theta } dy$$



and from here, it is not obvious to me that I should be able to interchange integration and differentiation and so I don't know how to proceed directly.



Thanks in advance.










share|cite|improve this question









$endgroup$




Suppose $f$ is a probability density on $y in mathbb{R}$ and $theta in mathbb{R}$ is some parameter.



Define $$S(theta) := log big( frac{partial f(y; theta)}{partial theta} big)$$



Then I'm trying to show that $E[S(theta)] = 0$. The derivation shown in my class is first noticing that, when $L(theta; y) = f(y; theta)$ and then seeing that:



begin{align}
0 &= frac{partial 1}{partial theta}dy \
&= frac{partial}{partial theta} int f(y; theta) dy \
&= int f(y; theta) frac{ frac{ partial L(theta; y) }{partial theta} }{ L(theta; y) } dy \
&=int f(y; theta) frac{ partial log(L(theta; y)) }{ partial theta} dy\
&= E[S(theta)]
end{align}



I think that my lack of intuition for this derivation may come from a weak background in analysis, that is, I'm not so intuitively clear on what conditions are required for interchanging differentiation and integration and so perhaps without this intuition I don't see why this result is obvious.



So, I was trying to gain more intuition by proceeding more directly, starting with $E[S(theta)]$ and then unpacking the result until I arrived at $$E[S(theta)] = int frac{ partial f(y; theta) }{partial theta } dy$$



and from here, it is not obvious to me that I should be able to interchange integration and differentiation and so I don't know how to proceed directly.



Thanks in advance.







statistics statistical-inference






share|cite|improve this question













share|cite|improve this question











share|cite|improve this question




share|cite|improve this question










asked Jan 25 at 18:42









libbylibby

1137




1137












  • $begingroup$
    The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
    $endgroup$
    – StubbornAtom
    Jan 25 at 18:48










  • $begingroup$
    Sure, but why do we need this assumption?
    $endgroup$
    – libby
    Jan 25 at 20:05


















  • $begingroup$
    The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
    $endgroup$
    – StubbornAtom
    Jan 25 at 18:48










  • $begingroup$
    Sure, but why do we need this assumption?
    $endgroup$
    – libby
    Jan 25 at 20:05
















$begingroup$
The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
$endgroup$
– StubbornAtom
Jan 25 at 18:48




$begingroup$
The fact that the integral and derivative can be interchanged is part of the assumptions ("regularity conditions").
$endgroup$
– StubbornAtom
Jan 25 at 18:48












$begingroup$
Sure, but why do we need this assumption?
$endgroup$
– libby
Jan 25 at 20:05




$begingroup$
Sure, but why do we need this assumption?
$endgroup$
– libby
Jan 25 at 20:05










0






active

oldest

votes











Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3087463%2fintuition-for-the-vanishing-expectation-of-score-es-theta-0%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























0






active

oldest

votes








0






active

oldest

votes









active

oldest

votes






active

oldest

votes
















draft saved

draft discarded




















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3087463%2fintuition-for-the-vanishing-expectation-of-score-es-theta-0%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

MongoDB - Not Authorized To Execute Command

How to fix TextFormField cause rebuild widget in Flutter

in spring boot 2.1 many test slices are not allowed anymore due to multiple @BootstrapWith