Open WenYanger opened 6 years ago
Same issue, any hints?
Same issue, any hints?
No idea, bro.
Same issue, any hints?
No idea, bro.
If this is still relevant to you. I found an issue with very long documents and words for my problem at hand. The current code is padding sequences and tokens based on the longest sequence and token within the current batch. So if you for example have a token with length 1000 in the batch, all tokens get padded to that size which can increase the necessity of memory allocation heavily.
One solution is to change the padding code in the package, or simpler, just do some pre-processing on your data to only create sequences of length X and tokens of length Y.
System information
Describe the problem
OOM(Out of memory) Error occured with a Warning :
However,a highly similar data could run on the same code. The size and format of data are all the same, looks like this:
I wonder which step in my code (or data) lead to such Warning, because another similar data haven't raised this Warning ~ T.T
Article on StackOverFlow said it is caused by TensorFlow function tf.gather(). Maybe it is the issue?
Source code / logs