Closed jaideep11061982 closed 1 year ago
Hi~
@zhiyuanyou
1) if I use a different image size say rectangular 512,768 . Which parameter do I have to adjust
as I get 0 for one of dimenstion of feat after using crop roi function
2) How do you chose the outstride, do I have to change it depending on the image resize shape I chose ?
3) Also I use resnet50, does the outstride value change in config as per backbone ?
4) I put some debug statements before crop roi , I get this for 512,768
feat torch.Size([1, 256, 128, 192]) tensor([[216., 497., 241., 527.],
[186., 403., 201., 425.],
[330., 498., 366., 518.],
[256., 245., 278., 279.]],
How does crop roi able to extract desired regions, as it positions overshot the feat shape on h,w dimensions
5) why do we divide the boxes by outstride in crop_roi function
boxes_scaled = boxes / out_stride
Hi~
@zhiyuanyou i figured out the logic. your crop roi and bbx transformation are based on the assumption that final fmap size will be /4 if there are changes into that then it disturbs BBx based crop computation in ROI
Yes. Cropping ROI is operated on feature map (whose size is _imagesize / _outstride), thus we need to divide _outstride first.
@zhiyuanyou is it possible to convert this into batch operation currently its a single image based both in training and inference. 1)can this converted to batch based operation. 2) Would increasing the batch size improve the gradients ?
3) I do not have DDP , i.e one gpu only can i comment this out in training
4) Now I am using backend as resnet50 and doing retraining using its pretrained wts but I get too less training loss ,is it expected
5) Did you train for 200 epochs ?