Open nonamestreet opened 7 years ago
Off the top of my head, it could be that you need to pass the train
argument into MRR score to exclude known interactions from ranking. However, I don't think this would account for the entirety of the difference.
I will have a look at reproducing this and get back to you in the next couple of days.
Thank your for the reply!
I just tried to put the train argument to MRR score, the performance is now at the 0.68-0.69 range which is quite close to the paper's result. But there is still a small gap left as you said.
How to get similar result to https://github.com/maciejkula/binge as in your paper?
The result is around 0.035 which is far from 0.07 in the paper. I was using the same hyper-parameters as in "movielens_1M_validation.log" in your binge repository, what am I missing?