Closed StefanIsSmart closed 9 months ago
Hi @StefanIsSmart,
Thanks for your interest in our work.
Let us know if you have other questions. Thanks.
Thank you for your reply! Also thank you for your work! It is really useful for me!
But I have one more question:
When I use the ComENet to do some tasks. For example, I use a learnable token to mask some position information as the input position Info. for ComENet. That means the input position Info. will require grads. The first iteration is okay, but after we did the backward() and step() , the nan will happen. This question only happened when I used ComENet, the other networks like EGNN did nothing wrong.
I don't know why, could you give me a favor?
Hi @StefanIsSmart,
Based on my experience, in most cases, nan happens when divided by zero. dist_ji here is used as the denominator in lines 375 and 383, so please try to add a small eps (e.g. 1e-6) to dist_ji to make sure the value is larger than 0.
Another difference between ComENet and other methods is that it selects some reference nodes based on distance values as shown here. I am not sure if this affects your method.
Let us know if you have other questions. Thanks.
Hi @StefanIsSmart,
Based on my experience, in most cases, nan happens when divided by zero. dist_ji here is used as the denominator in lines 375 and 383, so please try to add a small eps (e.g. 1e-6) to dist_ji to make sure the value is larger than 0.
Another difference between ComENet and other methods is that it selects some reference nodes based on distance values as shown here. I am not sure if this affects your method.
Let us know if you have other questions. Thanks.
I find the calculation of phi and tau will get the nan number in the backward() grad. The distance and the theta calculation is nothing wrong. I have no idea how to solve this bug....