Open MLDeS opened 1 year ago
We will release the source code soon.
@MLDeS
We unfreeze all layers by default and do not try to freeze some layers.
Nonetheless, we employ a learning rate decay strategy for mask-image-modeling (MIM) pre-trained models, a technique commonly used when fine-tuning MIM models. This strategy assigns a smaller learning rate to the shallower layers and a larger learning rate to the deeper ones, following the formula lr = base_lr * decay_rate ** (num_layers - layer_depth)
, where the decay_rate
is less than or equal to 1.
By adjusting the decay_rate, we can potentially achieve an effect similar to freezing some layers.
We have not yet evaluated the performance of frozen features within the DETR framework.
In a previous study(paper), we examined the use of frozen features for downstream dense tasks and compared different pre-training methods. We discovered that the performance of MIM frozen features was subpar, but this could be a result of poor classification. We would evaluate their localization performance later.
Hello,
Congratulations on the great work. I have some questions on the backbone used.
I am sorry if I ask any redundant question. It would be helpful to have some insights into these aspects.
Thanks a lot, again!