Open kikefdezl opened 1 year ago
Hi @kikefdezl, how many targets are there in a single picture of your dataset? This error occurred in the label assign process. The excessive number of gt caused it. Since the RBboxOverlaps2D
does not support CPU calculation now, it is recommended that you try using a smaller image size, e,g, 768x768 or 512x512
Hi @zytx121, thanks for the answer. That certainly explains it, as some of my images have a very high amount of boxes.
It was surprising to me, since I've been using ReDet for a long time now and memory usage never depended on target count. Does ReDet use a different method to assign Pred to GT?
Will this potentially be changed in the future for RTMDet?
@kikefdezl, DynamicSoftLabelAssigner
occupies more VRAM than MaxIoUAssigner.
You can reduce the topk
parameter to reduce the VRAM, which may damage some performance.
YES, we will consider your feedback in the next version.
Prerequisite
Task
I have modified the scripts/configs, or I'm working on my own tasks/models/datasets.
Branch
1.x branch https://github.com/open-mmlab/mmrotate/tree/1.x
Environment
sys.platform: linux Python: 3.8.10 (default, Nov 14 2022, 12:59:47) [GCC 9.4.0] CUDA available: True numpy_random_seed: 2147483648 GPU 0: NVIDIA A10G CUDA_HOME: /usr/local/cuda NVCC: Cuda compilation tools, release 11.6, V11.6.124 GCC: x86_64-linux-gnu-gcc (Ubuntu 9.4.0-1ubuntu1~20.04.1) 9.4.0 PyTorch: 1.13.1+cu117 PyTorch compiling details: PyTorch built with:
TorchVision: 0.14.1+cu117 OpenCV: 4.7.0 MMEngine: 0.3.0 MMRotate: 1.0.0rc1+5d0491c
Reproduces the problem - code sample
This is the full training config, based on
rotated_rtmdet_tiny-3x-dota.py
. It's the parsed version that is saved atwork-dir
when training:Reproduces the problem - command or script
Reproduces the problem - error message
Additional information
I believe this is the smallest available RTMDet-R model. I'm using an NVIDIA A10G with 24GB of VRAM and am running the dataloaders with batch size of 1, so I expected to be able to run this model.
I'm using a custom dataset, which has been pre-cropped to 1024x1024 images.
Is it normal for this model to use so much memory?