Open QAQEthan opened 3 years ago
I'm also experiencing CUDA out of memory issue with non-local block. I'm trying to use non-local block at the top of my network, which is for bbox regression conv head in faster r-cnn. Do you guys have any ideas to address this?
@Monkey-D-Luffy-star @vombategeht Hi~
The larger the size (height, width, depth) of feature maps is, the more memories the matrix multiplication will occupy.
When I encounter this problem,I will:
@AlexHex7 Thx, benefit a lot.
If non-local is applied to the low-level feature map, CUDA out of memory will happen.Is this due to the amount of memory required to compute the Attention matrix? Looking forward to your reply