Open sarmientoj24 opened 2 years ago
Hi @sarmientoj24, thanks for your interest in SparseInst. Have you load any pretrained weights?
Yes
On Mon, Aug 1, 2022, 17:37 Tianheng Cheng @.***> wrote:
Hi @sarmientoj24 https://github.com/sarmientoj24, thanks for your interest in SparseInst. Have you load any pretrained weights?
— Reply to this email directly, view it on GitHub https://github.com/hustvl/SparseInst/issues/66#issuecomment-1200961391, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACDL23Z2BH6QR4Y3GWFTRGTVW6LGZANCNFSM55BRMKUQ . You are receiving this because you were mentioned.Message ID: @.***>
Could you provide the log of the training process?
Could you provide the log of the training process?
Hello, I met the same problem as well when I try to change the optimizer from ADAMW to SGD.
I fell into this problem too. The cause is: SOLVER: AMP: ENABLED: true
Set it to False (use FP32) and the error disappears
I've tried to debug but wasn't able to fix it
I haven't loaded any pre-trained weights. The problem also exists.
Hi all, I've found that the sigmoid + norm
in the decoder will cause the NaN error when FP16 is enabled. In the latest update, we provide a special softmax version of the decoder to avoid numerical errors, and it supports FP16 better than the sigmoid + norm
. Sorry for the late reply and hope my suggestion can help you.
Hi all, I've found that the
sigmoid + norm
in the decoder will cause the NaN error when FP16 is enabled. In the latest update, we provide a special softmax version of the decoder to avoid numerical errors, and it supports FP16 better than thesigmoid + norm
. Sorry for the late reply and hope my suggestion can help you.
I got the same problem now. I am using pre-trained weights and trying to train R-50-vd-DCN model. Are there additional steps to use new softmax version?
It seems sigmoid + norm
is used by default. Adding MODEL.SPARSE_INST.DECODER.NAME GroupIAMSoftDecoder
to the command line solved the problem for me.
It seems
sigmoid + norm
is used by default. AddingMODEL.SPARSE_INST.DECODER.NAME GroupIAMSoftDecoder
to the command line solved the problem for me.
It still does not work...
When trying SparseInt with ViT, I get this error
Here's the config printed