Repository hosting code used to reproduce results in "Actions Speak Louder than Words: Trillion-Parameter Sequential Transducers for Generative Recommendations" (https://arxiv.org/abs/2402.17152).
Hi, great work!
I'm trying to reproduce the results on public datasets. However, I only found the training codes, where the model was evaluated on the eval set (or you don't use train/eval/test spilt, only train/test split?). I’d like to know if you partitioned the public dataset into a test set, and whether the results reported in the paper correspond to the test set or the eval set.
If I want to partition a test set, should I set ignore_last_n=0,1,2 when loading the test, eval and train dataset? For example:
Hi, great work! I'm trying to reproduce the results on public datasets. However, I only found the training codes, where the model was evaluated on the eval set (or you don't use train/eval/test spilt, only train/test split?). I’d like to know if you partitioned the public dataset into a test set, and whether the results reported in the paper correspond to the test set or the eval set. If I want to partition a test set, should I set
ignore_last_n=0,1,2
when loading the test, eval and train dataset? For example: