Closed yzhangcs closed 4 years ago
I am not sure why you encountered this error. In this implementation, we have some strategies to prevent from numerical overflow. For example, when calculating logdet(L), we first convert L to double precision. Second, during training, we skip mini-batches where NaN gradients occur.
For the CUDA error in your experiments, please print out L matrix to see if there are some illegal numbers such as inf or NaN.
Thanks for your reply. It seems to be the bug of the device (Titan V) itself. I try to run the model on GTX 2080Ti and it works.
Hi, thanks for your nice project. I encountered this error when I tried to train the
NeuroMST
model. Below are some error logs:I wonder what went wrong with me. By observing the outputs, I found that at the first epoch, the Laplacian matrix is in the following form:
where x corresponds a summation of exponents. Is it prone to cause the numerical overflow when calculating the determinants? I would be appreciated if given any suggestions.