[ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling"
Hello, your work is very meaningful and effective. While reviewing your code, I noticed a SparK.forward statement on line 156 in SparK/pretrain/main.py. I'm not quite familiar with this usage. Is it a typo or does it serve another purpose? I hope you can clarify this. Thank you.
Hello, your work is very meaningful and effective. While reviewing your code, I noticed a SparK.forward statement on line 156 in SparK/pretrain/main.py. I'm not quite familiar with this usage. Is it a typo or does it serve another purpose? I hope you can clarify this. Thank you.