This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows" on Object Detection and Instance Segmentation.
HI, am training on my personal dataset.
But am not using multiscale training and inference.
configs am using from here: '../base/models/cascade_mask_rcnn_swin_fpn.py',
but data loader and augmentation part am not using the: coco_instance.py
But i use the pretrained weights cascademask-rcnn without any issue..
so my questions are
1) those pretrained weights are not trained with MULTISCALE?
2) how can I use multiscale training? coz I dont see theres any part conbining prediction logits from multiscales
(I suppose the multiscale means from this paper? Hierarchical Multi-Scale Attention for Semantic Segmentation)
In addition, MULTISCALE in our experiments means inputs of different iterations or different GPUs may have different scales. These scales are sampled according to the rules shown in 2. In other words, in each iteration, on each GPU, there is only one scale of input.
HI, am training on my personal dataset. But am not using multiscale training and inference.
configs am using from here: '../base/models/cascade_mask_rcnn_swin_fpn.py', but data loader and augmentation part am not using the: coco_instance.py
But i use the pretrained weights cascademask-rcnn without any issue..
so my questions are 1) those pretrained weights are not trained with MULTISCALE? 2) how can I use multiscale training? coz I dont see theres any part conbining prediction logits from multiscales (I suppose the multiscale means from this paper? Hierarchical Multi-Scale Attention for Semantic Segmentation)
Thanks in advance :)