theFoxofSky / ddfnet

The official implementation of the CVPR2021 paper: Decoupled Dynamic Filter Networks
MIT License
214 stars 34 forks source link

About the dilation setting #22

Closed SudongCAI closed 3 years ago

SudongCAI commented 3 years ago

Hi author, thanks for your wonderful work. Today I noticed a problem in the dilation setting about the DDF model that "dilation" cannot set to 8 or more (like 16), otherwise the core will be dumped. I wonder how to solve this problem, even after checking the .cuda files... I will be very grateful if you can let me know how to solve this problem. Thank you very much.

theFoxofSky commented 3 years ago

I think this problem may be caused by the CUDA setting.

I also meet a similar problem when using very big batch sizes (for example, 384 or more for one GPU). I temporarily fixed the batch size issue in my private ddf repo by modifying Cuda files.

You may also need to modify the Cuda files if you want to use large dilation.

BTW, if you use large dilation, ddf will become very slow since the faster version cannot be used in the large dilation setting.

SudongCAI commented 3 years ago

Thank you so much for the reply, that's very kind. By the way, may I have another question: as for the 'mul_faster', I looked into the 'cuda' file and it seems no backward function was defined there. Is this faster version only for forward inference and cannot be independently applied in the training phase?

If 'mul_faster' is also working for the training phase, is there any difference on accuracy performances about the 'mul' and the 'mul_faster'?

theFoxofSky commented 3 years ago

No difference in the performance. I just optimized the forward part, I have not optimized the backward part yet. So the training will not be significantly faster. But the inference speed will be faster.

SudongCAI commented 3 years ago

Understood. Thank you so much.

jlhou commented 2 years ago

I would like to ask whether you have realized the code with larger "dilation" and could you share it with me.