Hi @hanoonaR , thanks for the interest of the work.
Yes, experiments are conducted and fine-tuned with the pre-trained model "swin UNETR/base". But note the number reported in the leaderboard is trained with additional training data. Please refer to Table 4 with a single fold, single model perforance.
Batch size is 4 with single GPU, the model is trained with 4 x 32G GPUs. Initial learning rate is set to 1e-4 x [number of GPUs] in the DDP training.
Hi, Thank you for sharing your great work. Could you please provide some clarity on the implementation details on the BTCV dataset for the Swin-UNETR.
1) Please clarify that the reported numbers on BTCV in the Self-Supervised Pre-Training of Swin Transformersfor 3D Medical Image Analysis paper are finetuned on the BTCV data from the pretrained model "Swin UNETR/Base".
2) Please share details on the batch size and learning rate used in finetuning for BTCV to get the reported numbers?
Thank you.