Closed jingwangsg closed 1 year ago
Hi, this is not deliberate: if I remember correctly this part of code is from the MDETR codebase which itself comes from the DETR codebase.
Hi, thank you for you reminder. I opened a new issue in original DETR repo. Not sure how this setting will affect the final performance, but it should be fixed if it's a bug indeed.
Seems it is solved https://github.com/facebookresearch/detr/issues/544.
Hi Antoine It seems that you set a different seed for each rank before building the model. This may lead to different parameter initialization for different duplicate on each rank. Is it a mistake or a deliberate design?
Here is a comment from pytorch lightning ddp advice