Closed XCRobert closed 5 years ago
You are right, stylization may change the bounding boxes by several pixels. We did not correct for that, but just used the original annotation files. Afaik, there is no algorithmic or general approach to solving that problem, since the bounding boxes vary depending on the style that is used and the style is randomly selected. As we applied stylization only to the training data (and used the concatenation of clean and stylized data most of the time) this didn't seem too much of an issue to us. Testing against clean aswell as corrupted data confirmed that performance was not hurt by the fact that the bounding boxes in the stylized dataset are not pixel perfect anymore.
As @Hvitgar already pointed out, potential alignment problems coming from the stylization were not an issue during training. We were worried though that it may cause problems when training for instance segmentation. We however found this not to be the case even when training only on stylized data.
For details compare Tables 2 & 4 in the paper: The performance drop going from Object Detection to Instance Segmentation is roughly the same for all models (~3% when evaluating on clean data) with no significant additional performance loss for the models trained only on stylized data.
One possible way is to stylize images with spatial control by using mask information. The performance may be improved.
Can you provide any resources to that method? Might be very interesting to have a look at that
Can you provide any resources to that method? Might be very interesting to have a look at that
How did you ensure the bounding box location of stylized images not to be changed ? Does it need additional annotation bboxes for stylized images.