Open TouqeerAhmad opened 4 years ago
I think the way Caffe is written is that each layer's Reshape()
method is called exactly once during layer construction (see the Setup()
method in layer.hpp
), and afterwards, the shape of the blobs remain the same throughout training/testing. So by default, you can't have different input sizes for different images.
With that said, it's possible to implement your own layer, which calls the layer's Reshape()
method inside each forward call. I am not 100% certain, but I have a feeling that it could negatively impact network speed.
Hi, Thank you for sharing the code and trained models! I have a question specific to the demo in your PyTorch implementation. As I understand, you are using a base_size = 512 to which any incoming image is resized regardless of the input dimensions while maintaining the original aspect ratio. Then the inference is run in a grid fashion on crops of 473x473.
My question is to why have a fixed crop_size or a base_size? There is no fully connected layer in the architecture so why we have this fixed size? Is this an arbitrary choice for numbers or there is a solid reason for it?
Thank you for your time! Best, Touqeer