Closed lantudou closed 3 years ago
Hi @lantudou CornerNet is very memory consuming, you can consider resize or modify batch, and then observe whether there is the same problem.
Hi @lantudou CornerNet is very memory consuming, you can consider resize or modify batch, and then observe whether there is the same problem.
Yeah..but i just want to know whether the evaluation process need extra cuda memory during the training. In other word, is it a normal situation or a code bug?
Hi @lantudou CornerNet is very memory consuming, you can consider resize or modify batch, and then observe whether there is the same problem.
Yeah..but i just want to know whether the evaluation process need extra cuda memory during the training. In other word, is it a normal situation or a code bug?
This is not easy to judge, but we have no problem training and verifying on v100.
@hhaAndroid v100 have large memory, of course no problem. Evaluation should be like detectron2, when there is not much memory left, the evaluation process should be transferred to cpu.
I tried run the mmdetection in colab environment and build up the lastest version of mmcv and mmdet using following command:
!pip install mmcv-full
The training process looks like ok but it is always killed due to cuda out of memory during the evaluation process. It really confused me beacause the memory cost of evaluation should be smaller than training process. Could you tell me where is the problem? Here is my training log as follows:
If it is necessary, i would like to share my colab code in here to reproduce the bug. Kind Regards!