tsinghua-fib-lab / CLSR

The official implementation of "Disentangling Long and Short-Term Interests for Recommendation" (WWW '22)
MIT License
70 stars 19 forks source link

Differences between the code and the paper #16

Closed zhuty16 closed 1 year ago

zhuty16 commented 1 year ago

Hello,

Thank you for providing the code. I wonder if there are some differences between the code and the description in the paper? For example, you use a user_short_embedding as the initial state of the interest_evolve GRU and also propose a discrepency loss between the long and short-term user embedding. But these are not mentioned in the paper. Moreover, the short-term query in the paper is the q_s^{u,t} but you use the concatenation of the short_term_intention and the target_item_embedding as the short-term query in the code.

Thank you!

DavyMorgan commented 1 year ago

Hello,

Thanks for your attention on our work. There are a few details missing in the paper due to space limit.

First, as introduced in the Remark of Section 3.2.4 in the paper, adding a discrepency loss is a traditional way to achieve disentanglement, however, it does not fit our case of disentangling long and short-term user interest since the two aspects can also overlap with each other to some extent. In our codes, we implement such discrepancy loss to investigate the effect of this traditional approach. Through experiments with different values of discrepancy_loss_weight, we find that the discrepancy loss brings no benefits. Specifically, the performance of setting discrepancy_loss_weight as a very small value of 0.01 is roughly the same as setting discrepancy_loss_weight as 0. Meanwhile, increasing discrepancy_loss_weight to 0.1 will bring performance drop. You can set discrepancy_loss_weight as other values to further investigate the effect of discrepancy loss.

Second, for the initial state of the interest_evolve GRU and the short-term query, these details are not introduced in the paper since they are not directly related to the main idea of self-supervised disentanglement. Please refer to our provided codes and we may update the paper on arxiv to include these details.

Thank you again!