Closed FightingEveryDay0 closed 1 year ago
Hi I wonder how did you manage to fix this problem? Because I'm facing the same issue too. Changing max_position_embedding
in config does not help.
Update after few minutes: I fixed this by changing the max_length_input
in primer_main.py
. Apparently dataloader does process and output 4097 tokens when the max is set to 4096. When I set the max_length_input
to 2048, the max I ended up getting is 2049.
When running the code on multi_news dataset, it raises error:
I found that there is only one dataset named "multi_news" and did anyone run the code
primer_main.py
on multi_news without the above problems? It's quite strange that the error only occur on the multi_news dataset in my experiment /(ㄒoㄒ)/~~ Thank you very much!