It states that the probability of a recommended item is the softmax value of the inner product (i.e., similarity) between the user representation and its embedding. However, the instruction to get softmax values is commented. Instead, values of the inner product are directly used to compute cross-entropy losses. Could you help me to understand the inconsistency between the paper and your implementation?
The statement in Improving Conversational Recommender Systems via Knowledge Graph based Semantic Fusion.
It states that the probability of a recommended item is the softmax value of the inner product (i.e., similarity) between the user representation and its embedding. However, the instruction to get softmax values is commented. Instead, values of the inner product are directly used to compute cross-entropy losses. Could you help me to understand the inconsistency between the paper and your implementation?
The statement in Improving Conversational Recommender Systems via Knowledge Graph based Semantic Fusion.