find partial derivative of the function












0












$begingroup$


I have this equation which is basically a maximum likelihood equation for EM-algorithm.
$$L(theta) = sum_{i=1}^n{ln{(sum_{j=1}^kw_jp_j(x_i;theta_j))}}$$
I'm trying to derive a partial derivative by $theta_j$ from it.
What I'm getting is:
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jover p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}$$
where $$p(x_i)=sum_{j=1}^kw_jp_j(x_i;theta_j)$$
This derivative should be equal to 0 and I can multiply it by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}p_j(x_i;theta_j)}{partial theta_j}=0$$
Which is almost as it should be. But the problem is that I somewhere miss the logarithm as it should look like this:
$$frac{partial sum_{i=1}^n g_{ij}ln p_j(x_i;theta_j)}{partial theta_j}=0$$
Where am I wrong?



$g_{ij}={w_jp_j(x_i;theta_j)over p(x_i)}$ which is obtained in the first step of the EM-algorithm.










share|cite|improve this question











$endgroup$












  • $begingroup$
    What are the $g_{i j}$ in the "answer"?
    $endgroup$
    – coffeemath
    Jan 10 at 18:04










  • $begingroup$
    @coffeemath added clarification into the question.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:09






  • 1




    $begingroup$
    You multiply and divide by $p_j(x_i;theta_j)$.
    $endgroup$
    – Math Lover
    Jan 10 at 18:13










  • $begingroup$
    @MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:21
















0












$begingroup$


I have this equation which is basically a maximum likelihood equation for EM-algorithm.
$$L(theta) = sum_{i=1}^n{ln{(sum_{j=1}^kw_jp_j(x_i;theta_j))}}$$
I'm trying to derive a partial derivative by $theta_j$ from it.
What I'm getting is:
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jover p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}$$
where $$p(x_i)=sum_{j=1}^kw_jp_j(x_i;theta_j)$$
This derivative should be equal to 0 and I can multiply it by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}p_j(x_i;theta_j)}{partial theta_j}=0$$
Which is almost as it should be. But the problem is that I somewhere miss the logarithm as it should look like this:
$$frac{partial sum_{i=1}^n g_{ij}ln p_j(x_i;theta_j)}{partial theta_j}=0$$
Where am I wrong?



$g_{ij}={w_jp_j(x_i;theta_j)over p(x_i)}$ which is obtained in the first step of the EM-algorithm.










share|cite|improve this question











$endgroup$












  • $begingroup$
    What are the $g_{i j}$ in the "answer"?
    $endgroup$
    – coffeemath
    Jan 10 at 18:04










  • $begingroup$
    @coffeemath added clarification into the question.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:09






  • 1




    $begingroup$
    You multiply and divide by $p_j(x_i;theta_j)$.
    $endgroup$
    – Math Lover
    Jan 10 at 18:13










  • $begingroup$
    @MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:21














0












0








0


1



$begingroup$


I have this equation which is basically a maximum likelihood equation for EM-algorithm.
$$L(theta) = sum_{i=1}^n{ln{(sum_{j=1}^kw_jp_j(x_i;theta_j))}}$$
I'm trying to derive a partial derivative by $theta_j$ from it.
What I'm getting is:
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jover p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}$$
where $$p(x_i)=sum_{j=1}^kw_jp_j(x_i;theta_j)$$
This derivative should be equal to 0 and I can multiply it by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}p_j(x_i;theta_j)}{partial theta_j}=0$$
Which is almost as it should be. But the problem is that I somewhere miss the logarithm as it should look like this:
$$frac{partial sum_{i=1}^n g_{ij}ln p_j(x_i;theta_j)}{partial theta_j}=0$$
Where am I wrong?



$g_{ij}={w_jp_j(x_i;theta_j)over p(x_i)}$ which is obtained in the first step of the EM-algorithm.










share|cite|improve this question











$endgroup$




I have this equation which is basically a maximum likelihood equation for EM-algorithm.
$$L(theta) = sum_{i=1}^n{ln{(sum_{j=1}^kw_jp_j(x_i;theta_j))}}$$
I'm trying to derive a partial derivative by $theta_j$ from it.
What I'm getting is:
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jover p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}$$
where $$p(x_i)=sum_{j=1}^kw_jp_j(x_i;theta_j)$$
This derivative should be equal to 0 and I can multiply it by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j} = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}p_j(x_i;theta_j)}{partial theta_j}=0$$
Which is almost as it should be. But the problem is that I somewhere miss the logarithm as it should look like this:
$$frac{partial sum_{i=1}^n g_{ij}ln p_j(x_i;theta_j)}{partial theta_j}=0$$
Where am I wrong?



$g_{ij}={w_jp_j(x_i;theta_j)over p(x_i)}$ which is obtained in the first step of the EM-algorithm.







derivatives maximum-likelihood log-likelihood






share|cite|improve this question















share|cite|improve this question













share|cite|improve this question




share|cite|improve this question








edited Jan 10 at 18:45









Kristian S. Jensen

54




54










asked Jan 10 at 17:53









s0nicYouths0nicYouth

34




34












  • $begingroup$
    What are the $g_{i j}$ in the "answer"?
    $endgroup$
    – coffeemath
    Jan 10 at 18:04










  • $begingroup$
    @coffeemath added clarification into the question.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:09






  • 1




    $begingroup$
    You multiply and divide by $p_j(x_i;theta_j)$.
    $endgroup$
    – Math Lover
    Jan 10 at 18:13










  • $begingroup$
    @MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:21


















  • $begingroup$
    What are the $g_{i j}$ in the "answer"?
    $endgroup$
    – coffeemath
    Jan 10 at 18:04










  • $begingroup$
    @coffeemath added clarification into the question.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:09






  • 1




    $begingroup$
    You multiply and divide by $p_j(x_i;theta_j)$.
    $endgroup$
    – Math Lover
    Jan 10 at 18:13










  • $begingroup$
    @MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
    $endgroup$
    – s0nicYouth
    Jan 10 at 18:21
















$begingroup$
What are the $g_{i j}$ in the "answer"?
$endgroup$
– coffeemath
Jan 10 at 18:04




$begingroup$
What are the $g_{i j}$ in the "answer"?
$endgroup$
– coffeemath
Jan 10 at 18:04












$begingroup$
@coffeemath added clarification into the question.
$endgroup$
– s0nicYouth
Jan 10 at 18:09




$begingroup$
@coffeemath added clarification into the question.
$endgroup$
– s0nicYouth
Jan 10 at 18:09




1




1




$begingroup$
You multiply and divide by $p_j(x_i;theta_j)$.
$endgroup$
– Math Lover
Jan 10 at 18:13




$begingroup$
You multiply and divide by $p_j(x_i;theta_j)$.
$endgroup$
– Math Lover
Jan 10 at 18:13












$begingroup$
@MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
$endgroup$
– s0nicYouth
Jan 10 at 18:21




$begingroup$
@MathLover no. Only multiply. Added clarification that this derivative should be equal to 0.
$endgroup$
– s0nicYouth
Jan 10 at 18:21










1 Answer
1






active

oldest

votes


















0












$begingroup$

It is not because your derivative is zero that you can multiply all the terms in your summation by different weights $p_j(x_i;theta_j)$ (they are different because they change with $x_i$). You should as mentioned by Math Lover, multiply and divide by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j}(theta) = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{1}{p_j(x_i;theta_j)} frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}log(p_j(x_i;theta_j))}{partial theta_j}=0.$$






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks, I got it. Stupid mistake.
    $endgroup$
    – s0nicYouth
    Jan 10 at 21:38











Your Answer





StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3068966%2ffind-partial-derivative-of-the-function%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























1 Answer
1






active

oldest

votes








1 Answer
1






active

oldest

votes









active

oldest

votes






active

oldest

votes









0












$begingroup$

It is not because your derivative is zero that you can multiply all the terms in your summation by different weights $p_j(x_i;theta_j)$ (they are different because they change with $x_i$). You should as mentioned by Math Lover, multiply and divide by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j}(theta) = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{1}{p_j(x_i;theta_j)} frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}log(p_j(x_i;theta_j))}{partial theta_j}=0.$$






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks, I got it. Stupid mistake.
    $endgroup$
    – s0nicYouth
    Jan 10 at 21:38
















0












$begingroup$

It is not because your derivative is zero that you can multiply all the terms in your summation by different weights $p_j(x_i;theta_j)$ (they are different because they change with $x_i$). You should as mentioned by Math Lover, multiply and divide by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j}(theta) = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{1}{p_j(x_i;theta_j)} frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}log(p_j(x_i;theta_j))}{partial theta_j}=0.$$






share|cite|improve this answer









$endgroup$













  • $begingroup$
    Thanks, I got it. Stupid mistake.
    $endgroup$
    – s0nicYouth
    Jan 10 at 21:38














0












0








0





$begingroup$

It is not because your derivative is zero that you can multiply all the terms in your summation by different weights $p_j(x_i;theta_j)$ (they are different because they change with $x_i$). You should as mentioned by Math Lover, multiply and divide by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j}(theta) = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{1}{p_j(x_i;theta_j)} frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}log(p_j(x_i;theta_j))}{partial theta_j}=0.$$






share|cite|improve this answer









$endgroup$



It is not because your derivative is zero that you can multiply all the terms in your summation by different weights $p_j(x_i;theta_j)$ (they are different because they change with $x_i$). You should as mentioned by Math Lover, multiply and divide by $p_j(x_i;theta_j)$ and get
$$frac{partial L}{partial theta_j}(theta) = sum_{i=1}^n{w_jp_j(x_i;theta_j)over p(x_i)}frac{1}{p_j(x_i;theta_j)} frac{partial p_j(x_i;theta_j)}{partial theta_j}=frac{partial sum_{i=1}^n g_{ij}log(p_j(x_i;theta_j))}{partial theta_j}=0.$$







share|cite|improve this answer












share|cite|improve this answer



share|cite|improve this answer










answered Jan 10 at 20:27









BertrandBertrand

16615




16615












  • $begingroup$
    Thanks, I got it. Stupid mistake.
    $endgroup$
    – s0nicYouth
    Jan 10 at 21:38


















  • $begingroup$
    Thanks, I got it. Stupid mistake.
    $endgroup$
    – s0nicYouth
    Jan 10 at 21:38
















$begingroup$
Thanks, I got it. Stupid mistake.
$endgroup$
– s0nicYouth
Jan 10 at 21:38




$begingroup$
Thanks, I got it. Stupid mistake.
$endgroup$
– s0nicYouth
Jan 10 at 21:38


















draft saved

draft discarded




















































Thanks for contributing an answer to Mathematics Stack Exchange!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


Use MathJax to format equations. MathJax reference.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3068966%2ffind-partial-derivative-of-the-function%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

'app-layout' is not a known element: how to share Component with different Modules

android studio warns about leanback feature tag usage required on manifest while using Unity exported app?

WPF add header to Image with URL pettitions [duplicate]