Pang-Yatian / Point-MAE

[ECCV2022] Masked Autoencoders for Point Cloud Self-supervised Learning
MIT License
448 stars 55 forks source link

Intuition behind the share-weighted learnable mask token #44

Closed nielsRocholl closed 6 months ago

nielsRocholl commented 8 months ago

Hi!

In your paper, you mention that:

For the embedding of each masked point patch, we replace it with a share-weighted learnable mask token

Could you provide some intuition behind why using a shared token for all masked tokens works? I had initially assumed that every patch would require a unique learnable masked token, but this of course, as proven by your work, is not the case.

Thanks!

Pang-Yatian commented 6 months ago

Well,.. it is a common practice to use the share-weighted learnable token. I think each patch using a unique token would cause overfitting problems.