Open bc-bytes opened 1 year ago
Greetings @bc-bytes, you have passed the pretty low image size (64x48) for the encoder. Please consider experimenting with input dimensions of 224x224 or 256x256. Additionally, make sure that your data format follows the structure of [batch_size, channel, height, width].
I need to train the models using patches of full images, which are 64 x 48 pixels.
Sorry! To meet the desired dimensions, you will need to change the architecture of the network. A size of 64x48 is small. Developing a segmentation model with this size may require designing a simple convolutional neural network with fewer layers. You may need to base it on UNet but with significantly smaller layers. You will need to create your UNet model for this task.
I am trying to train using the efficientnet_b0 backbone using an image size of 64x48, but get the following error:
I tried with and without pretrained model but got the same error each time.