XiaohangZhan / deocclusion

Code for our CVPR 2020 work.
Apache License 2.0
794 stars 104 forks source link

Are the bboxes of COCOA dataset incorrectly used in this code? #25

Closed Jiaoma closed 4 years ago

Jiaoma commented 4 years ago

Thanks for your code sharing. I'm a fresh man to this problem. When I looked into your ipython demo on COCOA dataset, I found that the amodal completion result of PCNet-M is always inside the bounding box provided by the COCOA dataset. However, the bounding box provided by the COCOA dataset seems only cover the modal annotations. Is it right? If it is true, I will feel confused about using modal bounding box to restrict the amodal completion area of PCNet-M. And I don't know whether it will bring any bad influence to your training stage. The figures below show an example I captured from demo_cocoa.ipynb. The image id is 2 (in code). The bounding boxes are: image The amodal completions are: image

XiaohangZhan commented 4 years ago

It is not as you described. For example, the left boundry of the completion of the dog in the image is out of the bbox. We enlarged the modal bbox to in the demo.

Jiaoma commented 4 years ago

Thank you for your reply. I reread the code and found that the code size = max([np.sqrt(bbox[2] * bbox[3] * self.config['enlarge_box']), bbox[2] * 1.1, bbox[3] * 1.1]) and function expand_bbox() are used for enlarging the origin bounding boxes. So the boxes used for amodal completion are larger than the original ones. I'm sorry for making the misleading descriptions.

I checked many amodal completion results of test images in COCOA dataset, most of them are perfectly matched the amodal gts, except some cases, like the one above.