Hi, thank you for implementing the awesome project!
I would like to use this project to the competition.
I have two questions about the preprocessing of input images during training and validation.
Describe the problem
At line 183 in train.py, an input image is divided by 255.0.
input_image = np.float32(input_image) / 255.0
However, for Resnet, the input image is pre-processed with the per-pixel mean subtracted. (https://arxiv.org/pdf/1512.03385.pdf).
Is there any reason to divide the pixel by 255?
At line 238 in train.py, an input image is cropped with crop_height and crop_width.
Information
Hi, thank you for implementing the awesome project! I would like to use this project to the competition. I have two questions about the preprocessing of input images during training and validation.
Describe the problem
At line 183 in train.py, an input image is divided by 255.0.
However, for Resnet, the input image is pre-processed with the per-pixel mean subtracted. (https://arxiv.org/pdf/1512.03385.pdf). Is there any reason to divide the pixel by 255?
At line 238 in train.py, an input image is cropped with crop_height and crop_width.
Is it "common" way to crop images at the test (validation) step? I thought at the test step, the image should be inputted into the network directly. (https://github.com/rishizek/tensorflow-deeplab-v3-plus/blob/master/inference.py)