This repository contains code for the paper "Margin Preserving Self-paced Contrastive Learning Towards Domain Adaptation for Medical Image Segmentation", published at IEEE JBHI 2022
Hi, I'm pretraining the deeplabv2 to generate prototypes on another dataset. I find that the prediction is down-sampled to 33 $\times$ 33 but not up-sampled. Why is that? Did you also down-sample labels? But it seems that you did not down-sample labels before calculating BCE and dice loss in train_UDA.py.
Note that I notice that there is a label_downsample function in train_UDA.py. But it is only used in the update_class_center_iter function, and that does not affect the shape of labels used for BCE and dice loss calculation.
Hi, I'm pretraining the deeplabv2 to generate prototypes on another dataset. I find that the prediction is down-sampled to 33 $\times$ 33 but not up-sampled. Why is that? Did you also down-sample labels? But it seems that you did not down-sample labels before calculating BCE and dice loss in train_UDA.py.
Note that I notice that there is a
label_downsample
function in train_UDA.py. But it is only used in theupdate_class_center_iter
function, and that does not affect the shape of labels used for BCE and dice loss calculation.Thanks.