Open volgachen opened 2 years ago
Sorry that we do not have the plan to release the code for pre-training, but it can be easily implemented by replacing the model construction function in the DeiT code with our model construction function.
Hope this can help you and feel free to ask anything if you have difficulties in implementing the pre-training code.
Thank you for your response.
I guess I should modify query_shape
into (1,1)
. Is there any other configs I should notice?
Taking fp-detr-base-in1k.py for example, there are several parts that should be modified in the config, as follows:
- The self-attn and the corresponding norm in encoder2 should be removed, and the operation order should be updated.
I suppose the self-attn
you mentioned in point 2 is actually prompt_self_attn?
Yes, that's right.
Thank you! It seems to be right now.
I find that there is a learning rate decay for sampling_offsets
in the training configuration for detection.
How do you handle with sampling_offsets
in the pretraining process?
We did not carefully tune the learning rate for sampling_offsets
and reference_points
during pre-training, and simply set their learning rate the same as other parameters in the transformer encoder. Tuning the learning rate may lead to better pre-training results, but we didn't try.
Excuse me, do you have any plan to release codes or instructions for pretraining?