Open fschvart opened 2 years ago
Hi,I meet same question。Have you found the reason and solve it,by the way,How did you change the batchsize? I didn't see the relevant settings
Hi,I meet same question。Have you found the reason and solve it,by the way,How did you change the batchsize? I didn't see the relevant settings
I left it for now. I don't know how to solve it. I found the batch size setting in the config file, or one of the parent config files and changed it there
Thank you for your timely reply. I plan to use the mask RCNN. The report needs 10.3G, but my 3060 and 12g are not enough. I also constructed a 6K data set myself. I was wondering whether some of his super parameters were designed for the coco data set. We couldn't make it clear, but the official didn't provide some instructions on how to use them. I'm sorry, but I looked carefully again, but I still couldn't find the batchsize. If it's convenient for you, I still hope you can tell me which file it is.
Hi,I meet same question。Have you found the reason and solve it,by the way,How did you change the batchsize? I didn't see the relevant settings
I left it for now. I don't know how to solve it. I found the batch size setting in the config file, or one of the parent config files and changed it there
just change value in config.default.py cannot solve it. because all code are install in system before. so if you want to change default value, you should use pip to reinstall detectron2. it is so bad for value as batch_size to change.
the way you should configure your batch size is for your own experiment, not globally.
To do it in a single run just add the last argument as here:
../../tools/lazyconfig_train_net.py --config-file configs/COCO/mask_rcnn_vitdet_b_100ep.py "dataloader.train.total_batch_size=1"
the way you should configure your batch size is for your own experiment, not globally. To do it in a single run just add the last argument as here:
thank you for your reply
Hi,I meet same question。Have you found the reason and solve it,by the way,How did you change the batchsize? I didn't see the relevant settings
I left it for now. I don't know how to solve it. I found the batch size setting in the config file, or one of the parent config files and changed it there
just change value in config.default.py cannot solve it. because all code are install in system before. so if you want to change default value, you should use pip to reinstall detectron2. it is so bad for value as batch_size to change.
In fact, I don't want to change the batch_ Size, but I always feel that the memory occupation in the table in the paper is not the same as that I understand, especially when it is 64, it occupies 10.9g, while when I am 2, the 12g 3060 is still insufficient
As kretes said, a batch size of 1 should fit in your GPU. From the information in their paper, the batch size they use for COCO finetuning is 64, distributed across 64 GPUs (1 image per GPU), and they use A100! I tried with a batch size of 2 on a 3090, but it takes 20GB so its borderline.
Hi, I have a similar problem, but I found that the ViTDet occupies more GPU memory than the plain ViT without the window partition. I am confused.
If you do not know the root cause of the problem, please post according to this template:
Instructions To Reproduce the Issue:
I'm trying to train an instance segmentation ViTDet model with a custom and relatively small dataset (6000 images of 640x480) I'm using Windows 10 and RTX 3090. I'm trying to train using the basic configuration ViTDet, ViT-B), which in theory, should take 12.3gb. My RTX 3090 has 24gb and I get a CUDA out of memory message. I reduced the batch size from 64 to 2, reduced the number of workers to 2 and used FP16 comperssion, none of it solved the issue. I checked the config.yaml in output and I see that these are my actual settings.
here's my environment:
PyTorch built with: