Zian-Xu / Swin-MAE

Pytorch implementation of Swin MAE https://arxiv.org/abs/2212.13805
62 stars 12 forks source link

Impelementation Details of finetung #7

Closed binyang97 closed 1 year ago

binyang97 commented 1 year ago

I'm currently using a similar architecture for the newtork as yours. I read the paper and if I understand correctly, are you loading the encoder weights into the decoder? Could you explain the reason behind this? It is little bit confusing to me.

Thank you for your time and looking forward to your reply! image

Zian-Xu commented 1 year ago

The swin Transformer blocks contained in the Swin-Unet's encoder and decoder are symmetric, thus compatible in the size of parameters. Experiments show that assigning the swin Transformer blocks in Swin-Unet's decoder the same weights as the encoder is better than random initialization. In the original Swin-Unet paper, the authors did transfer learning in the same way in their open-source code. I suggest you try two different types of training to verify that the results are improved.

------------------ 原始邮件 ------------------ 发件人: "Zian-Xu/Swin-MAE" @.>; 发送时间: 2023年8月18日(星期五) 下午5:17 @.>; @.***>; 主题: [Zian-Xu/Swin-MAE] Impelementation Details of finetung (Issue #7)

I'm currently using a similar architecture for the newtork as yours. I read the paper and if I understand correctly, are you loading the encoder weights into the decoder? Could you explain the reason behind this? It is little bit confusing to me.

Thank you for your time and looking forward to your reply!

— Reply to this email directly, view it on GitHub, or unsubscribe. You are receiving this because you are subscribed to this thread.Message ID: @.***>

binyang97 commented 1 year ago

Got it! Thank you so much for the reply