Closed rmathieu25 closed 3 years ago
Hi,
The OOM error you get is because your GPU memory cannot load the entire input mini-batch tensor. Here are some ways to get rid of the issue:
B
.L
.Reducing the number of cells may not help unless you only input several cells. That's quite weird. In your case, you could try to set L=50
or even a smaller number. E.g. if your maximum batch size for training is 512
, you need to make sure that B*L
is about 512
.
Thanks.
Hello, model.init_inference is very slow to run using the CPU version (but it is running) but I cannot get it to run by using the GPU version.
I get the following error:
I tried to reduce the batch size 64,32,16,8 but all failed. I am not running out of memory.
The is due to the size of the input data. When I reduce the number of cells in my data, it is working.
Thank you in advance.
Best regards