Open julled opened 2 years ago
Try to return zero loss for classification and regression when there is no gt in the image. You need to keep the keys in the loss dict same between normal images and no gt images.
Hi @RangiLyu and thanks for your reply!
i tried to set
filter_empty_gt=False
in my datasets to keep the examples without GT. Isnt this function supposed to guarantee this?
Try to return zero loss for classification and regression when there is no gt in the image. You need to keep the keys in the loss dict same between normal images and no gt images.
Why not set all loss functions, and return 0 in the loss function when there is no gt, for example-https://github.com/open-mmlab/mmdetection/blob/HEAD/mmdet/models/losses/smooth_l1_loss.py#L47
Hi @RangiLyu and thanks for your reply!
i tried to set
filter_empty_gt=False
in my datasets to keep the examples without GT. Isnt this function supposed to guarantee this?
I got the same question with you. Have you solved this problems by adding filter_empty_gt=False?
Hi @RangiLyu and thanks for your reply! i tried to set
filter_empty_gt=False
in my datasets to keep the examples without GT. Isnt this function supposed to guarantee this?I got the same question with you. Have you solved this problems by adding filter_empty_gt=False?
no , this didnt help. i think someone needs to implement the proposed changes by @RangiLyu or @wywywy01
I thought about this again and setting the loss to 0 would result in that there is no benefit from using the images without GT.
The Idea would be to use those empty images to reduce the false positives so we need some kind of loss.
I also met this problem after 2 epochs. Anyone can fix this problem?
I also met this problem after 2 epochs. Anyone can fix this problem?
could be that your data is corrupt, check this. this might have saved my problem
@RangiLyu,大佬,现在有解决的方法吗
Is there any updated on this?
@julled could you resolve this issue?
I ran into the same issue using Pytorch 1.13.0+cu117 and mmdetection 2.27.0.
Majority of images in my data set doesn't include any GT bboxes. The code runs fine with same data on single GPU. But got the error when running on 4 GPUs cluster.
I tried find_unused_parameters=True too, but ddin't help.
@amor-volastra not really, but as you mentioned it, i remember that my data also had a lot of examples without any GT bboxes. I actually resolved my problem my lowering the amount of examples without GT bboxes. So maybe, if this is possible for you, you could give this a try.
But better would be that mmdetection could actually handle this. For me without deeper knowledge, this is a hard to find bug.
Hello, I am a novice, my teacher asked me to try to calculate the loss of each gt on the picture separately, and then sum, but I encountered this problem when I was multi-card, I just added the traversal of gt_bbox and gt_list, and this problem occurred. Would you please ask me how to solve it
Hi,
i am using mmdet v2.20. I have a custom coco dataset with about 2k images with GroundTruth labels. I train a fasterrcnn and if i only use images with GT Labels everything learns smoothly. If i add about 0.5k additional images without GT Labels i get at random epochs the error 'loss log variables are different across GPUs!' and the training aborts. The message was built in to prevent GPU hanging via https://github.com/open-mmlab/mmsegmentation/pull/1035.
Can you give me a hint how to prevent this?
Here is my config: