Closed JiaxiangBU closed 4 years ago
Thanks for the feedback.
nitpick:
left * right
instead of left*right
.I can merge this once you resolve these two comments. Thanks.
and notes and space for relative probability weight left * right
, render LDA.Rmd
and use the relative path for the scriptLDA_functions.R
for reproducible purpose. @ethen8181 see 2dc23e2
Thanks. merged.
Hi, @ethen8181 I read your good intro for Gibbs Sampling. And I find a typo in one latex formula.
https://github.com/ethen8181/machine-learning/blob/1f71423da54bfde24de7528a3ef0f5c9e694f4b7/clustering_old/topic_model/LDA.Rmd#L149-L152
https://github.com/ethen8181/machine-learning/blob/1f71423da54bfde24de7528a3ef0f5c9e694f4b7/clustering_old/topic_model/LDA.Rmd#L198
Here is the thing, for the first iteration, you script does a random assignment for the first word in the first document. Here
left * right
with two values, the probability of topic 1 and that of topic 2. The sum of both is not equal to 1. It just displays the relative weight for this word between two topics. Thus, they need do a normalization.I double-check the reference you list at the end of the article. The author does use
\propto
to show relative weights for topics.Thus, I open a PR for further discussion.